KIPS Transactions on Computer and Communication Systems
/
v.6
no.1
/
pp.1-8
/
2017
Communications for mobile sink groups such as rescue teams or platoons bring about a new challenging issue for handling mobility in wireless sensor networks. To do this, many studies have been proposed to support mobile sink groups. When closely looking at mobile sink groups, they can be divided into (multiple) small groups according to the property of applications. For example, a platoon can be divided into multiple squads to carry out its mission in the battle field. However, the previous studies cannot efficiently support the division of mobile sink groups because they do not address three challenging issues engendered by the mobile sink group division. The first issue is to select a leader sink for a new small mobile sink group. The efficient data delivery from a source to small mobile sink groups is the second issue. Last, the third issue is to share data between leader sinks of small mobile sink groups. Thus, this paper proposes a routing protocol to efficiently support the division of mobile sink groups by solving the three challenging issues. For the first issue, the proposed protocol selects a leader sink of a new small mobile sink group which provide a minimum summation of the distance between the new leader sink and the previous leader sink and the distance from the new leader sink to all of its member sinks. For the efficient data delivery from a source to small mobile sink groups in the second issue, the proposed protocol determines the path to minimize the data dissemination distance from source to small mobile sink group by calculating with the location information of both the source and the leader sinks. With regard to the third issue, the proposed protocol exploits member sinks located among leader sinks to provide efficient data sharing among leaders sinks by considering the location information of member sinks. Simulation results verified that the proposed protocol is superior to the previous protocol in terms of the energy consumption.
The performance of the multiprocessor systems is limited by the several factors. The system performance is affected by the processor speed, memory delay, and interconnection network bandwidth/latency. By the evolution of semiconductor technology, off the shelf microprocessor speed breaks beyond GHz, and the processors can be scalable up to multiprocessor system by connecting through the interconnection networks. In this situation, the system performances are bound by the latencies and the bandwidth of the interconnection networks. SCI, Myrinet, and Gigabit Ethernet are widely adopted as a high-speed interconnection network links for the high performance cluster systems. Performance improvement of the interconnection network can be achieved by the bandwidth extension and the latency minimization. Speed up of the operation clock speed is a simple way to accomplish the bandwidth and latency betterment, while its physical distance makes the difficulties to attain the high frequency clock. Hence the system performance and scalability suffered from the interconnection network limitation. Duplicating the link of the interconnection network is one of the solutions to resolve the bottleneck of the scalable systems. Dual-ring SCI link structure is an example of the interconnection network improvement. In this paper, I propose a network topology and a transaction path algorism, which optimize the latency and the efficiency under the duplicated links. By the simulation results, the proposed structure shows 1.05 to 1.11 times better latency, and exhibits 1.42 to 2.1 times faster execution compared to the dual ring systems.
Monitoring is used to see if a real-time system provides a service on time. Generally, monitoring for real-time focuses on investigating the current status of a real-time system. To support a stable performance of a real-time system, it should have not only a function to see the current status of real-time process but also a function to predict executions of real-time processes, however. The legacy prediction model has some limitation to apply it to a real-time monitoring. First, it performs a static prediction after a real-time process finished. Second, it needs a statistical pre-analysis before a prediction. Third, transition probability and data about clustering is not based on the current data. We propose the execution prediction model based on learning algorithm to solve these problems and apply it to real-time monitoring. This model gets rid of unnecessary pre-processing and supports a precise prediction based on current data. In addition, this supports multi-level prediction by a trend analysis of past execution data. Most of all, We designed the model to support dynamic prediction which is performed within a real-time process' execution. The results from some experiments show that the judgment accuracy is greater than 80% if the size of a training set is set to over 10, and, in the case of the multi-level prediction, that the prediction difference of the multi-level prediction is minimized if the number of execution is bigger than the size of a training set. The execution prediction model proposed in this model has some limitation that the model used the most simplest learning algorithm and that it didn't consider the multi-regional space model managing CPU, memory and I/O data. The execution prediction model based on a learning algorithm proposed in this paper is used in some areas related to real-time monitoring and control.
The existing management job of research expenses has been divided into three parts: budget planning, budget draw-up, and exact settlement of budget. However, it caused some problems. Under this current circumstance it is required to obtain research expenses steadily, to operate efficiently and to use them clearly to solve such problems. As a result of a study on data warehouse development process of existing system integration company (Inmon, IBM) to reflect current trend described above, data warehouse development process of Inmon uses systematic and gradual access as a classical development cycle method. It causes overlap and feedback to the previous step in the process of each step Is requested. And another problem that it is difficult to toil what function refers and corrects data because functions and data are separated during performing development process at data warehouse development process of IBM is caused. Integrated management data warehouse development process of research expenses in the enterprise environment which applies UML at planning and analysis step, design step and implement and test step is suggested in this paper. Information retrieval agent uses existing budget plan DB, budget draw-up DB and budget settlement DB to find out information that a user wants to know. Information retrieval agent collects and saves information at integration database and information integration agent extracts, transports, transforms and loads the data. Information integration agent reduces a user's efforts to access to a number of information sources and check each of them. It also screens out data that a user may not need. As a result, integrated management data warehouse development process of research expenses in the enterprise environment reflects a user's requirements as much as possible and provides various types of information to make a decision which is needed to establish the policy of research expense management. It helps an end user approach his/her desired analysis information quickly and get various data from the comprehensive viewpoint rather than the fragmentary viewpoint. Furthermore, as it integrated three systems into one, it is possible to share data, to integrate the system, to reduce operating expenses and to simplify supporting environment for the decision making.
Since 2011, DCF(Discounted Cash Flow) method has been used initiatively for valuating R&D technology assets in the agricultural food industry and recently technology valuation based on royalties comparison among technology transfer transactions has been also carried out in parallel when evaluating the technology assets such as new seed development technologies. Since the DCF method which has been known until now has many input variables to be estimated, sophisticated estimation has been demanded at the time of technology valuation. In addition, considering more similar trading cases when applying sales transaction comparison or industry norm method based on information of technology transfer royalty, it is an important issue that should be taken into account in the same way in the Agri-Food industry. The main input variables used for technology valuation in the Agri-Food industry are life cycle of technology asset, the financial information related to the Agri-Food industry, discount rate, and technology contribution rate. The latest infrastructure building and data updating related to technology valuation has been carried out on a regular basis in the evaluation organization of the Agri-Food segment. This study verifies the key variables that give the most important impact on the results for the existing technology valuation in the Agri-Food industry and clarifies the difference between the existing valuation result and the outcome by referring the support information that is derived through the latest input information applied in DCF method. In addition, while presenting the scheme to complement fragment information which the latest input data just influence result of technology valuation, we tried to perform comparative analysis between the existing valuation results and the evaluated outcome after the latest of reference data for making a decision the input values to be estimated in DCF. To perform these analyzes, it was first selected the representative cases evaluated past in the Agri-Food industry, applied a sensitivity analysis for input variables based on these selected cases, and then executed a simulation analysis utilizing the key input variables derived from sensitivity analysis. The results of this study is to provide the information which there are the need for modernization of the data related to the input variables that are utilized during valuating technology assets in the Agri-Food sector and for building the infrastructure of the key input variables in DCF. Therefore it is expected to provide more fruitful information about the results of valuation.
Alien turtles belonging to Genus Trachemys have been designated as Invasive Alien Species since 2001 and their import has been banned in Korea. However, current status of import and distribution of the other alien turtles have not been reported. In this study, we aimed to investigate the taxa of alien turtles introduced into Korea, to assess their potential risks to the natural ecosystems and to suggest the future management directions of them in Korea. We identified 73 species of alien turtles belonging to 9 families. Since 2008, more than 6,000 kg of turtles have been imported annually and widely distributed through the pet shops, traditional markets and individual transactions. From the survey of natural habitats, we found that 8 species belonging to 3 families including Chrysemys picta, Pseudemys concinna, P. nelsoni, P. peninsularis, P. rubriventris, Mauremys sinensis, Macrochelys temminckii and Trachemys scripta have inhabited in 12 study sites. Out of 73 alien turtles, the potential adverse impacts of 13 species to ecosystems are serious when we considered status of designation of invasive alien species in other countries. For the management of alien turtles, it is required to register alien turtles in the import list and share general information such as import purpose, distribution and management condition among relevant authorities. The breeders and distributors must be obliged to identify turtles and to record management. The government must check transfer and migration of turtles periodically to prevent their introduction and spread into natural environments. The change of alien turtle populations in natural habitats should be monitored and their management plan should be developed to control the alien turtles in areas where the impacts are significant.
Journal of the Korea Academia-Industrial cooperation Society
/
v.19
no.2
/
pp.279-292
/
2018
Unlike the global art market which experienced rapid recovery from the impacts of the Global Financial Crisis in 2008, the Korean art market has not yet fully recovered. The gallery-oriented distribution system, vulnerable primary art market functions, and the market structure centered on a small number of collectors make it difficult for young and medium artists to enter the market and, as a result, deepen the economic polarization of artists. In addition, the high price of art works limits market participation by restricting the general public. This study began with the idea that the interest of the public in the art market as well as their participation in the market are urgent. To this end, we noted that public awareness of art transactions can be a starting point for improving the constitution of the fragile art market, focusing on the 'Artist-centered Art Fair' rather than existing art fairs. To examine the contribution of such an art fair to the popularization of the art market, we analyzed the case of the 'Visual Artist Market (VAM)' project of the Korea Arts Management Service. Results found that the 'Artist-centered Art Fair' focuses on providing opportunities for market entry to young and medium artists rather than on the interests of distributors, and promotes the popularization of the art market by promoting low-priced works to the general public. Also, the 'Artist-centered Art Fair' seems to play a primary role in the public sector to foster solid groups of artists as well as to establish healty distribution networks of Korean Art market. However, in the long run, it is necessary to promote sustainable development of the 'Artist-centered Art Fair' through indirect support, such as the provision of a publicity platform or consumer finance support, rather than direct support.
One of the major problems in the area of data mining is the size of the data, as most data set has huge volume these days. Streams of data are normally accumulated into data storages or databases. Transactions in internet, mobile devices and ubiquitous environment produce streams of data continuously. Some data set are just buried un-used inside huge data storage due to its huge size. Some data set is quickly lost as soon as it is created as it is not saved due to many reasons. How to use this large size data and to use data on stream efficiently are challenging questions in the study of data mining. Stream data is a data set that is accumulated to the data storage from a data source continuously. The size of this data set, in many cases, becomes increasingly large over time. To mine information from this massive data, it takes too many resources such as storage, money and time. These unique characteristics of the stream data make it difficult and expensive to store all the stream data sets accumulated over time. Otherwise, if one uses only recent or partial of data to mine information or pattern, there can be losses of valuable information, which can be useful. To avoid these problems, this study suggests a method efficiently accumulates information or patterns in the form of rule set over time. A rule set is mined from a data set in stream and this rule set is accumulated into a master rule set storage, which is also a model for real-time decision making. One of the main advantages of this method is that it takes much smaller storage space compared to the traditional method, which saves the whole data set. Another advantage of using this method is that the accumulated rule set is used as a prediction model. Prompt response to the request from users is possible anytime as the rule set is ready anytime to be used to make decisions. This makes real-time decision making possible, which is the greatest advantage of this method. Based on theories of ensemble approaches, combination of many different models can produce better prediction model in performance. The consolidated rule set actually covers all the data set while the traditional sampling approach only covers part of the whole data set. This study uses a stock market data that has a heterogeneous data set as the characteristic of data varies over time. The indexes in stock market data can fluctuate in different situations whenever there is an event influencing the stock market index. Therefore the variance of the values in each variable is large compared to that of the homogeneous data set. Prediction with heterogeneous data set is naturally much more difficult, compared to that of homogeneous data set as it is more difficult to predict in unpredictable situation. This study tests two general mining approaches and compare prediction performances of these two suggested methods with the method we suggest in this study. The first approach is inducing a rule set from the recent data set to predict new data set. The seocnd one is inducing a rule set from all the data which have been accumulated from the beginning every time one has to predict new data set. We found neither of these two is as good as the method of accumulated rule set in its performance. Furthermore, the study shows experiments with different prediction models. The first approach is building a prediction model only with more important rule sets and the second approach is the method using all the rule sets by assigning weights on the rules based on their performance. The second approach shows better performance compared to the first one. The experiments also show that the suggested method in this study can be an efficient approach for mining information and pattern with stream data. This method has a limitation of bounding its application to stock market data. More dynamic real-time steam data set is desirable for the application of this method. There is also another problem in this study. When the number of rules is increasing over time, it has to manage special rules such as redundant rules or conflicting rules efficiently.
Today, IT organizations perform projects with vision related to marketing and financial profit. The objective of realizing the vision is to improve the project performing ability in terms of QCD. Organizations have made a lot of efforts to achieve this objective through process improvement. Large companies such as IBM, Ford, and GE have made over $80\%$ of success through business process re-engineering using information technology instead of business improvement effect by computers. It is important to collect, analyze and manage the data on performed projects to achieve the objective, but quantitative measurement is difficult as software is invisible and the effect and efficiency caused by process change are not visibly identified. Therefore, it is not easy to extract the strategy of improvement. This paper measures and analyzes the project performance, focusing on organizations' external effectiveness and internal efficiency (Qualify, Delivery, Cycle time, and Waste). Based on the measured project performance scores, an OT (Opportunity Tree) model was designed for optimizing the project performance. The process of design is as follows. First, meta data are derived from projects and analyzed by quantitative GQM(Goal-Question-Metric) questionnaire. Then, the project performance model is designed with the data obtained from the quantitative GQM questionnaire and organization's performance score for each area is calculated. The value is revised by integrating the measured scores by area vision weights from all stakeholders (CEO, middle-class managers, developer, investor, and custom). Through this, routes for improvement are presented and an optimized improvement method is suggested. Existing methods to improve software process have been highly effective in division of processes' but somewhat unsatisfactory in structural function to develop and systemically manage strategies by applying the processes to Projects. The proposed OT model provides a solution to this problem. The OT model is useful to provide an optimal improvement method in line with organization's goals and can reduce risks which may occur in the course of improving process if it is applied with proposed methods. In addition, satisfaction about the improvement strategy can be improved by obtaining input about vision weight from all stakeholders through the qualitative questionnaire and by reflecting it to the calculation. The OT is also useful to optimize the expansion of market and financial performance by controlling the ability of Quality, Delivery, Cycle time, and Waste.
The Transactions of the Korea Information Processing Society
/
v.5
no.10
/
pp.2575-2583
/
1998
The productionof the highly relible softwae systems and theirs performance evaluation hae become important interests in the software industry. The software evaluation has been mainly carried out in ternns of both reliability and performance of software system. Software reliability is the probability that no software error occurs for a fixed time interval during software testing phase. These theoretical software reliability models are sometimes unsuitable for the practical testing phase in which a software error at a certain testing stage occurs by causes of the imperfect debugging, abnornal software correction, and so on. Such a certatin software testing stage needs to be considered as an outlying stage. And we can assume that the software reliability does not improve by means of muisance factor in this outlying testing stage. In this paper, we discuss Bavesian software reliability growth modeling and estimation procedure in the presence of an imidentitied outlying software testing stage by the modification of Jehnski Moranda. Also we derive the Bayes estimaters of the software reliability panmeters by the assumption of prior information under the squared error los function. In addition, we evaluate the proposed software reliability growth model with an unidentified outlying stage in an exchangeable model according to the values of nuisance paramether using the accuracy, bias, trend, noise metries as the quantilative evaluation criteria through the compater simulation.
본 웹사이트에 게시된 이메일 주소가 전자우편 수집 프로그램이나
그 밖의 기술적 장치를 이용하여 무단으로 수집되는 것을 거부하며,
이를 위반시 정보통신망법에 의해 형사 처벌됨을 유념하시기 바랍니다.
[게시일 2004년 10월 1일]
이용약관
제 1 장 총칙
제 1 조 (목적)
이 이용약관은 KoreaScience 홈페이지(이하 “당 사이트”)에서 제공하는 인터넷 서비스(이하 '서비스')의 가입조건 및 이용에 관한 제반 사항과 기타 필요한 사항을 구체적으로 규정함을 목적으로 합니다.
제 2 조 (용어의 정의)
① "이용자"라 함은 당 사이트에 접속하여 이 약관에 따라 당 사이트가 제공하는 서비스를 받는 회원 및 비회원을
말합니다.
② "회원"이라 함은 서비스를 이용하기 위하여 당 사이트에 개인정보를 제공하여 아이디(ID)와 비밀번호를 부여
받은 자를 말합니다.
③ "회원 아이디(ID)"라 함은 회원의 식별 및 서비스 이용을 위하여 자신이 선정한 문자 및 숫자의 조합을
말합니다.
④ "비밀번호(패스워드)"라 함은 회원이 자신의 비밀보호를 위하여 선정한 문자 및 숫자의 조합을 말합니다.
제 3 조 (이용약관의 효력 및 변경)
① 이 약관은 당 사이트에 게시하거나 기타의 방법으로 회원에게 공지함으로써 효력이 발생합니다.
② 당 사이트는 이 약관을 개정할 경우에 적용일자 및 개정사유를 명시하여 현행 약관과 함께 당 사이트의
초기화면에 그 적용일자 7일 이전부터 적용일자 전일까지 공지합니다. 다만, 회원에게 불리하게 약관내용을
변경하는 경우에는 최소한 30일 이상의 사전 유예기간을 두고 공지합니다. 이 경우 당 사이트는 개정 전
내용과 개정 후 내용을 명확하게 비교하여 이용자가 알기 쉽도록 표시합니다.
제 4 조(약관 외 준칙)
① 이 약관은 당 사이트가 제공하는 서비스에 관한 이용안내와 함께 적용됩니다.
② 이 약관에 명시되지 아니한 사항은 관계법령의 규정이 적용됩니다.
제 2 장 이용계약의 체결
제 5 조 (이용계약의 성립 등)
① 이용계약은 이용고객이 당 사이트가 정한 약관에 「동의합니다」를 선택하고, 당 사이트가 정한
온라인신청양식을 작성하여 서비스 이용을 신청한 후, 당 사이트가 이를 승낙함으로써 성립합니다.
② 제1항의 승낙은 당 사이트가 제공하는 과학기술정보검색, 맞춤정보, 서지정보 등 다른 서비스의 이용승낙을
포함합니다.
제 6 조 (회원가입)
서비스를 이용하고자 하는 고객은 당 사이트에서 정한 회원가입양식에 개인정보를 기재하여 가입을 하여야 합니다.
제 7 조 (개인정보의 보호 및 사용)
당 사이트는 관계법령이 정하는 바에 따라 회원 등록정보를 포함한 회원의 개인정보를 보호하기 위해 노력합니다. 회원 개인정보의 보호 및 사용에 대해서는 관련법령 및 당 사이트의 개인정보 보호정책이 적용됩니다.
제 8 조 (이용 신청의 승낙과 제한)
① 당 사이트는 제6조의 규정에 의한 이용신청고객에 대하여 서비스 이용을 승낙합니다.
② 당 사이트는 아래사항에 해당하는 경우에 대해서 승낙하지 아니 합니다.
- 이용계약 신청서의 내용을 허위로 기재한 경우
- 기타 규정한 제반사항을 위반하며 신청하는 경우
제 9 조 (회원 ID 부여 및 변경 등)
① 당 사이트는 이용고객에 대하여 약관에 정하는 바에 따라 자신이 선정한 회원 ID를 부여합니다.
② 회원 ID는 원칙적으로 변경이 불가하며 부득이한 사유로 인하여 변경 하고자 하는 경우에는 해당 ID를
해지하고 재가입해야 합니다.
③ 기타 회원 개인정보 관리 및 변경 등에 관한 사항은 서비스별 안내에 정하는 바에 의합니다.
제 3 장 계약 당사자의 의무
제 10 조 (KISTI의 의무)
① 당 사이트는 이용고객이 희망한 서비스 제공 개시일에 특별한 사정이 없는 한 서비스를 이용할 수 있도록
하여야 합니다.
② 당 사이트는 개인정보 보호를 위해 보안시스템을 구축하며 개인정보 보호정책을 공시하고 준수합니다.
③ 당 사이트는 회원으로부터 제기되는 의견이나 불만이 정당하다고 객관적으로 인정될 경우에는 적절한 절차를
거쳐 즉시 처리하여야 합니다. 다만, 즉시 처리가 곤란한 경우는 회원에게 그 사유와 처리일정을 통보하여야
합니다.
제 11 조 (회원의 의무)
① 이용자는 회원가입 신청 또는 회원정보 변경 시 실명으로 모든 사항을 사실에 근거하여 작성하여야 하며,
허위 또는 타인의 정보를 등록할 경우 일체의 권리를 주장할 수 없습니다.
② 당 사이트가 관계법령 및 개인정보 보호정책에 의거하여 그 책임을 지는 경우를 제외하고 회원에게 부여된
ID의 비밀번호 관리소홀, 부정사용에 의하여 발생하는 모든 결과에 대한 책임은 회원에게 있습니다.
③ 회원은 당 사이트 및 제 3자의 지적 재산권을 침해해서는 안 됩니다.
제 4 장 서비스의 이용
제 12 조 (서비스 이용 시간)
① 서비스 이용은 당 사이트의 업무상 또는 기술상 특별한 지장이 없는 한 연중무휴, 1일 24시간 운영을
원칙으로 합니다. 단, 당 사이트는 시스템 정기점검, 증설 및 교체를 위해 당 사이트가 정한 날이나 시간에
서비스를 일시 중단할 수 있으며, 예정되어 있는 작업으로 인한 서비스 일시중단은 당 사이트 홈페이지를
통해 사전에 공지합니다.
② 당 사이트는 서비스를 특정범위로 분할하여 각 범위별로 이용가능시간을 별도로 지정할 수 있습니다. 다만
이 경우 그 내용을 공지합니다.
제 13 조 (홈페이지 저작권)
① NDSL에서 제공하는 모든 저작물의 저작권은 원저작자에게 있으며, KISTI는 복제/배포/전송권을 확보하고
있습니다.
② NDSL에서 제공하는 콘텐츠를 상업적 및 기타 영리목적으로 복제/배포/전송할 경우 사전에 KISTI의 허락을
받아야 합니다.
③ NDSL에서 제공하는 콘텐츠를 보도, 비평, 교육, 연구 등을 위하여 정당한 범위 안에서 공정한 관행에
합치되게 인용할 수 있습니다.
④ NDSL에서 제공하는 콘텐츠를 무단 복제, 전송, 배포 기타 저작권법에 위반되는 방법으로 이용할 경우
저작권법 제136조에 따라 5년 이하의 징역 또는 5천만 원 이하의 벌금에 처해질 수 있습니다.
제 14 조 (유료서비스)
① 당 사이트 및 협력기관이 정한 유료서비스(원문복사 등)는 별도로 정해진 바에 따르며, 변경사항은 시행 전에
당 사이트 홈페이지를 통하여 회원에게 공지합니다.
② 유료서비스를 이용하려는 회원은 정해진 요금체계에 따라 요금을 납부해야 합니다.
제 5 장 계약 해지 및 이용 제한
제 15 조 (계약 해지)
회원이 이용계약을 해지하고자 하는 때에는 [가입해지] 메뉴를 이용해 직접 해지해야 합니다.
제 16 조 (서비스 이용제한)
① 당 사이트는 회원이 서비스 이용내용에 있어서 본 약관 제 11조 내용을 위반하거나, 다음 각 호에 해당하는
경우 서비스 이용을 제한할 수 있습니다.
- 2년 이상 서비스를 이용한 적이 없는 경우
- 기타 정상적인 서비스 운영에 방해가 될 경우
② 상기 이용제한 규정에 따라 서비스를 이용하는 회원에게 서비스 이용에 대하여 별도 공지 없이 서비스 이용의
일시정지, 이용계약 해지 할 수 있습니다.
제 17 조 (전자우편주소 수집 금지)
회원은 전자우편주소 추출기 등을 이용하여 전자우편주소를 수집 또는 제3자에게 제공할 수 없습니다.
제 6 장 손해배상 및 기타사항
제 18 조 (손해배상)
당 사이트는 무료로 제공되는 서비스와 관련하여 회원에게 어떠한 손해가 발생하더라도 당 사이트가 고의 또는 과실로 인한 손해발생을 제외하고는 이에 대하여 책임을 부담하지 아니합니다.
제 19 조 (관할 법원)
서비스 이용으로 발생한 분쟁에 대해 소송이 제기되는 경우 민사 소송법상의 관할 법원에 제기합니다.
[부 칙]
1. (시행일) 이 약관은 2016년 9월 5일부터 적용되며, 종전 약관은 본 약관으로 대체되며, 개정된 약관의 적용일 이전 가입자도 개정된 약관의 적용을 받습니다.