Recently, XML has been emerging as a standard for storing and exchanging of data for various distributed applications based on the Internet. Since there are increasing demands to store and manage XML documents, a lot of research works are going on this area to develop new took and techniques based on the XHL. However, most of the researches are concentrated on mapping techniques based on instance or DTD, and the main focus is on structural transformation. Current trend of research is toward the usage of XML documents based on XML schema, and demands not only conversion of structure but also preservation of the semantic constraints. This paper sets up the using of DIDL standing on the basis of XML schema from MPEG-21 as an application domain, and proposes the mapping model that can preserve semantic constraints in addition. We expand previous research techniques in the preprocessing step for the specific domain, and then, apply various new mapping methods in the postprocessing step. We present and discuss the system architecture for implementation, and introduce the algorithms and present implementation environment and semantic extension methodology in detail. Finally we show actual table and query processing based on our proposal.
This paper presents a technique to implement embedded systems using interrupts of the one-chip microcontroller with many peripherals based on a multiple finite state machines model. The multiple finite state machine model utilizes the structure of FSMD used for hardware design and the features of flow control by interrupts. The main finite state machine corresponds to the main program and the sub-state machines corresponds to the interrupt subroutines. Therefore, interrupts from the peripherals can be processed immediately in the sub-state machines. The request and reply variables are used to interface between the finite state machines. Additional operating system is not necessary for the context switching between the main state machine and the sub-state machine, because the flow-control caused by interrupt can be replaced with the switching. An embedded system modeled on multiple finite state machine with ASM charts can be easily implemented by the conversion of ASM charts into C-language programs. This implementation technique can be easily adopted to the hardware oriented embedded systems because of the detail description of the model and the fast response to the interrupt events in the sub-state machine.
In global supply chain, the scheduling problems of large products such as ships, airplanes, space shuttles, assembled constructions, and/or automobiles are complicated by nature. New scheduling systems are often developed in order to reduce inherent computational complexity. As a result, a problem can be decomposed into small sub-problems, problems that contain independently small scheduling systems integrating into the initial problem. As one of the authors experienced, DAS (Daewoo Shipbuilding Scheduling System) has adopted a two-layered hierarchical architecture. In the hierarchical architecture, individual scheduling systems composed of a high-level dock scheduler, DAS-ERECT and low-level assembly plant schedulers, DAS-PBS, DAS-3DS, DAS-NPS, and DAS-A7 try to search the best schedules under their own constraints. Moreover, the steep growth of communication technology and logistics enables it to introduce distributed multi-nation production plants by which different parts are produced by designated plants. Therefore vertical and lateral coordination among decomposed scheduling systems is necessary. No standard coordination mechanism of multiple scheduling systems exists, even though there are various scheduling systems existing in the area of scheduling research. Previous research regarding the coordination mechanism has mainly focused on external conversation without capacity model. Prior research has heavily focuses on agent-based coordination in the area of agent research. Yet, no scheduling domain has been developed. Previous research regarding the agent-based scheduling has paid its ample attention to internal coordination of scheduling process, a process that has not been efficient. In this study, we suggest a general framework for agent-based coordination of multiple scheduling systems in global supply chain. The purpose of this study was to design a standard coordination mechanism. To do so, we first define an individual scheduling agent responsible for their own plants and a meta-level coordination agent involved with each individual scheduling agent. We then suggest variables and values describing the individual scheduling agent and meta-level coordination agent. These variables and values are represented by Backus-Naur Form. Second, we suggest scheduling agent communication protocols for each scheduling agent topology classified into the system architectures, existence or nonexistence of coordinator, and directions of coordination. If there was a coordinating agent, an individual scheduling agent could communicate with another individual agent indirectly through the coordinator. On the other hand, if there was not any coordinating agent existing, an individual scheduling agent should communicate with another individual agent directly. To apply agent communication language specifically to the scheduling coordination domain, we had to additionally define an inner language, a language that suitably expresses scheduling coordination. A scheduling agent communication language is devised for the communication among agents independent of domain. We adopt three message layers which are ACL layer, scheduling coordination layer, and industry-specific layer. The ACL layer is a domain independent outer language layer. The scheduling coordination layer has terms necessary for scheduling coordination. The industry-specific layer expresses the industry specification. Third, in order to improve the efficiency of communication among scheduling agents and avoid possible infinite loops, we suggest a look-ahead load balancing model which supports to monitor participating agents and to analyze the status of the agents. To build the look-ahead load balancing model, the status of participating agents should be monitored. Most of all, the amount of sharing information should be considered. If complete information is collected, updating and maintenance cost of sharing information will be increasing although the frequency of communication will be decreasing. Therefore the level of detail and updating period of sharing information should be decided contingently. By means of this standard coordination mechanism, we can easily model coordination processes of multiple scheduling systems into supply chain. Finally, we apply this mechanism to shipbuilding domain and develop a prototype system which consists of a dock-scheduling agent, four assembly- plant-scheduling agents, and a meta-level coordination agent. A series of experiments using the real world data are used to empirically examine this mechanism. The results of this study show that the effect of agent-based platform on coordinated scheduling is evident in terms of the number of tardy jobs, tardiness, and makespan.
Our study is focused on a multiple-agent system to provide efficient collaborative work by automating the conference calling process with the help of intelligent agents. Automating the meeting scheduling requires a careful consideration of the individual official schedule as well as the privacy and personal preferences. Therefore, the automation of conference calling needs the distributed processing task where a separate calendar management process is associated for increasing the reliability and inherent parallelism. This paper describes in detail the design and implementation issues of a multiple-agent system for conference calling that allows the convener and participants to minimize their efforts in creating a meeting. Our system is based on the client-sewer model. In the sewer side, a scheduling agent, a negotiating agent, a personal information managing agent, a group information managing agent, a session managing agent, and a coordinating agent are operating. In the client side, an interface agent, a media agent, and a collaborating agent are operating. Agents use a standardized knowledge manipulation language to communicate amongst themselves. Communicating through a standardized knowledge manipulation language allows the system to overcome heterogeneity which is one of the most important problems in communication among agents for distributed collaborative computing. The agents of our system propose the dates on which as many participants as possible are available to attend the conference using the forward chaining algorithm and the back propagation network algorithm.
밀티빔 음향 측심기 (Multibeam Echo Sounder)는 탐사선에 수직방향으로 해저면을 주사(Swath)하여, 한번의 송수신(Ping)으로 다중의 빔자료를 얻을 수 있는 측심기로, 해저면에 반사되어 되돌아오는 음파의 음압을 기록하고, 사이드 스캔 소나 자료도 동시에 취득하는 기능을 가지고 있으므로, 측심된 해저 지형(Bathymetry)과 해저 지형을 덮고 있는 해저면의 퇴적 상황(Sediment Environment)도 동시에 얻을 수 있는 다목적 측심기이다. 본 논문에서는 L3사의 Sea Beam 2100 멀티빔 음향 측심기를 통해 얻은 자료를 처리하여, 3차원 공간 데이터인 DEM(Digital Elevation Model)을 생성하고, VRML을 이용한 웹상에서의 해저 지형 가시화를 통해, 세계 어느 곳에서나 웹을 통하여 쉽게 정보를 공유할 수 있는 3차원 해저 지리 정보 시스템의 구현을 목적으로 한다. 멀티빔 음향 측심기를 통해 얻어진 자료는 항해 자료 보정, 음속 보정, 빔 좌표 계산과 분리, 오측심 자료 제거, 조석 보정 등의 단계를 거쳐 측심자료의 정확도 및 신뢰도를 높이는 과정을 거치게 된다. 보정된 멀티빔 음향 측심자료는 무작위 점 사상(Point Topology)으로 산재 되어 있는 빔 자료를 임의의 단위영역으로 변환하는 과정을 거쳐야 하는데, 이 과정을 격자화라고 한다. 자료의 격자화를 통해 3차원 공강 데이터인 DEM 파일을 제작하고, 이 DEM 파일과 음압 영상을 이용해 웹상에서의 3차원 해저 지형의 가시화를 실현한다. 웹상에서의 3차원 지형 가시화에서 방대한 양의 지형 데이터는 데이터 전송 시간과 렌더링 시간에 치명적인 문제이다. 따라서, 렌더링 시간과 데이터 전송 시간을 단축시키기 위한, 지형 자료의 LOD(Level of Detail)를 통해, VRML을 이용한 보다 효과적인 웹상에서의 3차원 해저 지형의 가시화를 실현한다.면 기업은 고객으로 공간적인 제약으로 인한 불신을 불식시키는 신뢰감을 주게 된다. 이러한 고객서비스 향상과 물류비용 절감은 사이버 쇼핑몰이 전국 어디서나 우리의 안방에서 자연스럽게 점할 수 있는 상황을 만들 것이다.SP가 도입되어, 설계업무를 지원하기위한 기본적인 시스템 구조를 구상하게 된다. 이와 함께 IT Model을 구성하게 되는데, 객체지향적 접근 방법으로 Model을 생성하고 UML(Unified Modeling Language)을 Tool로 사용한다. 단계 4)는 Software Engineering 관점으로 접근한다. 이는 최종산물이라고 볼 수 있는 설계업무 지원 시스템을 Design하는 과정으로, 시스템에 사용될 데이터를 Design하는 과정과, 데이터를 기반으로 한 기능을 Design하는 과정으로 나눈다. 이를 통해 생성된 Model에 따라 최종적으로 Coding을 통하여 실제 시스템을 구축하게 된다.the making. program and policy decision making, The objectives of the study are to develop the methodology of modeling the socioeconomic evaluation, and build up the practical socioeconomic evaluation model of the HAN projects including scientific and technological effects. Since the HAN projects consists of 18 subprograms, it is difficult In evaluate all the subprograms simultaneously. Despite, each program is being performed under the category of HAN projects, so the common soci
Electronics cultural atlas is a typical methodology for the Digital Humanities. Since electronic cultural atlas could visualize various cultural information based on electronic atlas, that is an information visualization, users are able to make intuitive understanding on specific areas. Therefore, it can be effectively utilized in Area Studies and also helps users to understand comprehensively information on historic event in particular area with time information, because electronic cultural atlas represents a particular subject and time information with geographical information based on map. In other words, electronic cultural atlas may be considered as a specialized system of Digital Humanities for studying the Humanities and Area Studies. In this paper, we design and implement mediterranean electronic cultural atlas(MECA) for researchers of the Mediterranean area that has cultural hybridity formed the exchange of various aspects such as civilization, religion, race and language. In detail, a 'Digital Humanities Research Support System' is constructed to visualize research outcomes related to the Mediterranean area on Electronic Cultural Atlas and to use for researches.
The purpose of the study was to development an information delivery system for the home nursing service, to demonstrate and to evaluate the efficiency of it. The period of research conduct was from September 1996 to August 31, 1997. At the 1st stage to achieve the purpose, Firstly Assessment tool for the patients with cerebral vascular disease who have the first priority of HNS among the patients with various health problems at home was developed through literature review. Secondly, after identification of patient nursing problem by the home care nurse with the assessment tool, the patient's classification system developed by Park (1988) that was 128 nursing activities under 6 categories was used to identify the home care nurse's activities of the patient with CAV at home. The research team had several workshops with 5 clinical nurse experts to refine it. At last 110 nursing activities under 11 categories for the patients with CVA were derived. At the second stage, algorithms were developed to connect 110 nursing activities with the patient nursing problems identified by assessment tool. The computerizing process of the algorithms is as follows:
Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.
The wall shear stress in the vicinity of end-to end anastomoses under steady flow conditions was measured using a flush-mounted hot-film anemometer(FMHFA) probe. The experimental measurements were in good agreement with numerical results except in flow with low Reynolds numbers. The wall shear stress increased proximal to the anastomosis in flow from the Penrose tubing (simulating an artery) to the PTFE: graft. In flow from the PTFE graft to the Penrose tubing, low wall shear stress was observed distal to the anastomosis. Abnormal distributions of wall shear stress in the vicinity of the anastomosis, resulting from the compliance mismatch between the graft and the host artery, might be an important factor of ANFH formation and the graft failure. The present study suggests a correlation between regions of the low wall shear stress and the development of anastomotic neointimal fibrous hyperplasia(ANPH) in end-to-end anastomoses. 30523 T00401030523 ^x Air pressure decay(APD) rate and ultrafiltration rate(UFR) tests were performed on new and saline rinsed dialyzers as well as those roused in patients several times. C-DAK 4000 (Cordis Dow) and CF IS-11 (Baxter Travenol) reused dialyzers obtained from the dialysis clinic were used in the present study. The new dialyzers exhibited a relatively flat APD, whereas saline rinsed and reused dialyzers showed considerable amount of decay. C-DAH dialyzers had a larger APD(11.70
The wall shear stress in the vicinity of end-to end anastomoses under steady flow conditions was measured using a flush-mounted hot-film anemometer(FMHFA) probe. The experimental measurements were in good agreement with numerical results except in flow with low Reynolds numbers. The wall shear stress increased proximal to the anastomosis in flow from the Penrose tubing (simulating an artery) to the PTFE: graft. In flow from the PTFE graft to the Penrose tubing, low wall shear stress was observed distal to the anastomosis. Abnormal distributions of wall shear stress in the vicinity of the anastomosis, resulting from the compliance mismatch between the graft and the host artery, might be an important factor of ANFH formation and the graft failure. The present study suggests a correlation between regions of the low wall shear stress and the development of anastomotic neointimal fibrous hyperplasia(ANPH) in end-to-end anastomoses. 30523 T00401030523 ^x Air pressure decay(APD) rate and ultrafiltration rate(UFR) tests were performed on new and saline rinsed dialyzers as well as those roused in patients several times. C-DAK 4000 (Cordis Dow) and CF IS-11 (Baxter Travenol) reused dialyzers obtained from the dialysis clinic were used in the present study. The new dialyzers exhibited a relatively flat APD, whereas saline rinsed and reused dialyzers showed considerable amount of decay. C-DAH dialyzers had a larger APD(11.70