• Title/Summary/Keyword: Local Computing

Search Result 533, Processing Time 0.024 seconds

A Study of Perceived Value and Intention to Use for Car Sharing Service : Based on User Experiences Serviced by Seoul Car Sharing (차량공유 서비스에 대한 지각된 가치와 이용의향에 관한 연구 : 서울시 나눔카 서비스 이용자를 중심으로)

  • Park, Keon Chul;Song, In-Kuk
    • Journal of Internet Computing and Services
    • /
    • v.20 no.2
    • /
    • pp.109-118
    • /
    • 2019
  • The purpose of this study is to deliver both market-practical and civil-centric political implication for sharing economy by investigating the nature of consumer-adoption for car-sharing service. With the global interest and market proliferation of the sharing economy, various service models for sharing idle resources have also been released in Korea. Particularly, in case of car sharing service, public - private partnership projects are spreading rapidly in various local governments including Seoul, along with the growing demand for alternative transportation system centering on the urban area. This study conducted an empirical study on the process of accepting the car sharing service by analyzing the data collected from users of the car sharing service "Sharing Car(Nanum Car)" of Seoul Metropolitan Government. A survey was conducted on 281 users in their twenties who are in the age of main use among the experienced users of the "Sharing Car(NaNum)" residing in Seoul. The result of analysis on the relationship between these users' perceived value and intention to use the vehicle sharing service would provide implications for establishing consumer(citizen)-centeric policies as well as market implications.

Shared Key and Public Key based Mobile Agent Authentication Scheme supporting Multiple Domain in Home Network Environments (홈 네트워크 환경에서 다중 도메인을 지원하는 공유키 및 공개키 기반의 이동 에이전트 인증 기법)

  • 김재곤;김구수;엄영익
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.14 no.5
    • /
    • pp.109-119
    • /
    • 2004
  • The home network environment can be defined as a network environment, connecting digital home devices such as computer systems, digital appliances, and mobile devices. In this kind of home network environments, there will be numerous local/remote interactions to monitor and control the home network devices and the home gateway. Such an environment may result in communication bottleneck. By applying the mobile agents that can migrate among the computing devices autonomously and work on behalf of the user, remote interactions and network traffics can be reduced enormously. The mobile agent authentication is necessary to apply mobile agent concept to the home network environments, as a prerequisite technology for authorization or access control to the home network devices and resources. The existing mobile agent systems have mainly used the public key based authentication scheme, which is not suitable to the home network environments, composed of digital devices of limited computation capability. In this paper, we propose a shared key based mobile agent authentication scheme for single home domain and expand the scheme to multiple domain environments with the public key based authentication scheme. Application of the shared key encryption scheme to the single domain mobile agent authentication enables to authenticate the mobile agent with less overhead than the public key based authentication scheme.

Past, Present and Future of Geospatial Scheme based on Topo-Climatic Model and Digital Climate Map (소기후모형과 전자기후도를 기반으로 한 지리공간 도식의 과거, 현재 그리고 미래)

  • Kim, Dae-Jun
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.23 no.4
    • /
    • pp.268-279
    • /
    • 2021
  • The geospatial schemes based on topo-climatology have been developed to produce digital climate maps at a site-specific scale. Their development processes are reviewed here to derive the needs for new schemes in the future. Agricultural and forestry villages in Korea are characterized by complexity and diversity in topography, which results in considerably large spatial variations in weather and climate over a small area. Hence, the data collected at a mesoscale through the Automated Synoptic Observing System (ASOS) operated by the Korea Meteorological Administration (KMA) are of limited use. The geospatial schemes have been developed to estimate climate conditions at a local scale, e.g., 30 m, lowering the barriers to deal with the processes associated with production in agricultural and forestry industries. Rapid enhancement of computing technologies allows for near real-time production of climate information at a high-resolution even in small catchment areas and the application to future climate change scenarios. Recent establishment of the early warning service for agricultural weather disasters can provide growth progress and disaster forecasts for cultivated crops on a farm basis. The early warning system is being expanded worldwide, requiring further advancement in geospatial schemes and digital climate mapping.

A Study on the remote acuisition of HejHome Air Cloud artifacts (스마트 홈 헤이 홈 Air의 클라우드 아티팩트 원격 수집 방안 연구)

  • Kim, Ju-eun;Seo, Seung-hee;Cha, Hae-seong;Kim, Yeok;Lee, Chang-hoon
    • Journal of Internet Computing and Services
    • /
    • v.23 no.5
    • /
    • pp.69-78
    • /
    • 2022
  • As the use of Internet of Things (IoT) devices has expanded, digital forensics coverage of the National Police Agency has expanded to smart home areas. Accordingly, most of the existing studies conducted to acquire smart home platform data were mainly conducted to analyze local data of mobile devices and analyze network perspectives. However, meaningful data for evidence analysis is mainly stored on cloud storage on smart home platforms. Therefore, in this paper, we study how to acquire stored in the cloud in a Hey Home Air environment by extracting accessToken of user accounts through a cookie database of browsers such as Microsoft Edge, Google Chrome, Mozilia Firefox, and Opera, which are recorded on a PC when users use the Hey Home app-based "Hey Home Square" service. In this paper, the it was configured with smart temperature and humidity sensors, smart door sensors, and smart motion sensors, and artifacts such as temperature and humidity data by date and place, device list used, and motion detection records were collected. Information such as temperature and humidity at the time of the incident can be seen from the results of the artifact analysis and can be used in the forensic investigation process. In addition, the cloud data acquisition method using OpenAPI proposed in this paper excludes the possibility of modulation during the data collection process and uses the API method, so it follows the principle of integrity and reproducibility, which are the principles of digital forensics.

New Soil Classification System Using Cone Penetration Test (콘관입시험결과를 이용한 새로운 흙분류 방법의 개발)

  • Kim, Chan-Hong;Im, Jong-Chul;Kim, Young-Sang;Joo, No-Ah
    • Journal of the Korean Geotechnical Society
    • /
    • v.24 no.10
    • /
    • pp.57-70
    • /
    • 2008
  • The advantage of piezocone penetration test is a guarantee of continuous data, which is a source of reliable interpretation of target soil layer. Many researches have been carried out f3r several decades and several classification charts have been developed to classify in-situ soil from the cone penetration test result. Since most present classification charts or methods were developed based on the data which were compiled over the world except Korea, they should be verified to be feasible for Korean soil. Furthermore, sometimes their charts provide different soil classification results according to the different input parameters. However, unfortunately, revision of those charts is quite difficult or almost impossible. In this research a new soil classification model is proposed by using fuzzy C-mean clustering and neuro-fuzzy theory based on the 5371 CPT results and soil logging results compiled from 17 local sites around Korea. Proposed neuro-fuzzy soil classification model was verified by comparing the classification results f3r new data, which were not used during learning process of neuro-fuzzy model, with real soil log. Efficiency of proposed neuro-fuzzy model was compared with other soft computing classification models and Robertson method for new data.

Comparison and Analysis of Information Exchange Distributed Algorithm Performance Based on a Circular-Based Ship Collision Avoidance Model (원형 기반 선박 충돌 피항 모델에 기반한 정보 교환 분산알고리즘 성능 비교 분석)

  • Donggyun Kim
    • Journal of Navigation and Port Research
    • /
    • v.47 no.6
    • /
    • pp.401-409
    • /
    • 2023
  • This study compared and analyzed the performance of a distributed area search algorithm and a distributed probability search algorithm based on information exchange between ships. The distributed algorithm is a method that can search for an optimal avoidance route based on information exchange between ships. In the distributed area search algorithm, only a ship with the maximum cost reduction among neighboring ships has priority, so the next expected location can be changed. The distributed stochastic search algorithm allows a non-optimal value to be searched with a certain probability so that a new value can be searched. A circular-based ship collision avoidance model was used for the ship-to-ship collision avoidance experiment. The experimental method simulated the distributed area search algorithm and the distributed stochastic search algorithm while increasing the number of ships from 2 to 50 that were the same distance from the center of the circle. The calculation time required for each algorithm, sailing distance, and number of message exchanges were compared and analyzed. As a result of the experiment, the DSSA(Distributed Stochastic Search Algorithm) recorded a 25%calculation time, 88% navigation distance, and 84% of number of message exchange rate compared to DLSA.

GWB: An integrated software system for Managing and Analyzing Genomic Sequences (GWB: 유전자 서열 데이터의 관리와 분석을 위한 통합 소프트웨어 시스템)

  • Kim In-Cheol;Jin Hoon
    • Journal of Internet Computing and Services
    • /
    • v.5 no.5
    • /
    • pp.1-15
    • /
    • 2004
  • In this paper, we explain the design and implementation of GWB(Gene WorkBench), which is a web-based, integrated system for efficiently managing and analyzing genomic sequences, Most existing software systems handling genomic sequences rarely provide both managing facilities and analyzing facilities. The analysis programs also tend to be unit programs that include just single or some part of the required functions. Moreover, these programs are widely distributed over Internet and require different execution environments. As lots of manual and conversion works are required for using these programs together, many life science researchers suffer great inconveniences. in order to overcome the problems of existing systems and provide a more convenient one for helping genomic researches in effective ways, this paper integrates both managing facilities and analyzing facilities into a single system called GWB. Most important issues regarding the design of GWB are how to integrate many different analysis programs into a single software system, and how to provide data or databases of different formats required to run these programs. In order to address these issues, GWB integrates different analysis programs byusing common input/output interfaces called wrappers, suggests a common format of genomic sequence data, organizes local databases consisting of a relational database and an indexed sequential file, and provides facilities for converting data among several well-known different formats and exporting local databases into XML files.

  • PDF

Overlay Multicast Network for IPTV Service using Bandwidth Adaptive Distributed Streaming Scheme (대역폭 적응형 분산 스트리밍 기법을 이용한 IPTV 서비스용 오버레이 멀티캐스트 네트워크)

  • Park, Eun-Yong;Liu, Jing;Han, Sun-Young;Kim, Chin-Chol;Kang, Sang-Ug
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.12
    • /
    • pp.1141-1153
    • /
    • 2010
  • This paper introduces ONLIS(Overlay Multicast Network for Live IPTV Service), a novel overlay multicast network optimized to deliver live broadcast IPTV stream. We analyzed IPTV reference model of ITU-T IPTV standardization group in terms of network and stream delivery from the source networks to the customer networks. Based on the analysis, we divide IPTV reference model into 3 networks; source network, core network and access network, ION(Infrastructure-based Overlay Multicast Network) is employed for the source and core networks and PON(P2P-based Overlay Multicast Network) is applied to the access networks. ION provides an efficient, reliable and stable stream distribution with very negligible delay while PON provides bandwidth efficient and cost effective streaming with a little tolerable delay. The most important challenge in live P2P streaming is to reduce end-to-end delay without sacrificing stream quality. Actually, there is always a trade-off between delay & stream quality in conventional live P2P streaming system. To solve this problem, we propose two approaches. Firstly, we propose DSPT(Distributed Streaming P2P Tree) which takes advantage of combinational overlay multicasting. In DSPT, a peer doesn't fully rely on SP(Supplying Peer) to get the live stream, but it cooperates with its local ANR(Access Network Relay) to reduce delay and improve stream quality. When RP detects bandwidth drop in SP, it immediately switches the connection from SP to ANR and continues to receive stream without any packet loss. DSPT uses distributed P2P streaming technique to let the peer share the stream to the extent of its available bandwidth. This means, if RP can't receive the whole stream from SP due to lack of SP's uploading bandwidth, then it receives only partial stream from SP and the rest from the ANR. The proposed distributed P2P streaming improves P2P networking efficiency.

Development of a Remote Multi-Task Debugger for Qplus-T RTOS (Qplus-T RTOS를 위한 원격 멀티 태스크 디버거의 개발)

  • 이광용;김흥남
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.9 no.4
    • /
    • pp.393-409
    • /
    • 2003
  • In this paper, we present a multi-task debugging environment for Qplus-T embedded-system such as internet information appliances. We will propose the structure and functions of a remote multi-task debugging environment supporting environment effective ross-development. And, we are going enhance the communication architecture between the host and target system to provide more efficient cross-development environment. The remote development toolset called Q+Esto consists to several independent support tools: an interactive shell, a remote debugger, a resource monitor, a target manager and a debug agent. Excepting a debug agent, all these support tools reside on the host systems. Using the remote multi-task debugger on the host, the developer can spawn and debug tasks on the target run-time system. It can also be attached to already-running tasks spawned from the application or from interactive shell. Application code can be viewed as C/C++ source, or as assembly-level code. It incorporates a variety of display windows for source, registers, local/global variables, stack frame, memory, event traces and so on. The target manager implements common functions that are shared by Q+Esto tools, e.g., the host-target communication, object file loading, and management of target-resident host tool´s memory pool and target system´s symbol-table, and so on. These functions are called OPEn C APIs and they greatly improve the extensibility of the Q+Esto Toolset. The Q+Esto target manager is responsible for communicating between host and target system. Also, there exist a counterpart on the target system communicating with the host target manager, which is called debug agent. Debug agent is a daemon task on real-time operating systems in the target system. It gets debugging requests from the host tools including debugger via target manager, interprets the requests, executes them and sends the results to the host.

Hardware Approach to Fuzzy Inference―ASIC and RISC―

  • Watanabe, Hiroyuki
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1993.06a
    • /
    • pp.975-976
    • /
    • 1993
  • This talk presents the overview of the author's research and development activities on fuzzy inference hardware. We involved it with two distinct approaches. The first approach is to use application specific integrated circuits (ASIC) technology. The fuzzy inference method is directly implemented in silicon. The second approach, which is in its preliminary stage, is to use more conventional microprocessor architecture. Here, we use a quantitative technique used by designer of reduced instruction set computer (RISC) to modify an architecture of a microprocessor. In the ASIC approach, we implemented the most widely used fuzzy inference mechanism directly on silicon. The mechanism is beaded on a max-min compositional rule of inference, and Mandami's method of fuzzy implication. The two VLSI fuzzy inference chips are designed, fabricated, and fully tested. Both used a full-custom CMOS technology. The second and more claborate chip was designed at the University of North Carolina(U C) in cooperation with MCNC. Both VLSI chips had muliple datapaths for rule digital fuzzy inference chips had multiple datapaths for rule evaluation, and they executed multiple fuzzy if-then rules in parallel. The AT & T chip is the first digital fuzzy inference chip in the world. It ran with a 20 MHz clock cycle and achieved an approximately 80.000 Fuzzy Logical inferences Per Second (FLIPS). It stored and executed 16 fuzzy if-then rules. Since it was designed as a proof of concept prototype chip, it had minimal amount of peripheral logic for system integration. UNC/MCNC chip consists of 688,131 transistors of which 476,160 are used for RAM memory. It ran with a 10 MHz clock cycle. The chip has a 3-staged pipeline and initiates a computation of new inference every 64 cycle. This chip achieved an approximately 160,000 FLIPS. The new architecture have the following important improvements from the AT & T chip: Programmable rule set memory (RAM). On-chip fuzzification operation by a table lookup method. On-chip defuzzification operation by a centroid method. Reconfigurable architecture for processing two rule formats. RAM/datapath redundancy for higher yield It can store and execute 51 if-then rule of the following format: IF A and B and C and D Then Do E, and Then Do F. With this format, the chip takes four inputs and produces two outputs. By software reconfiguration, it can store and execute 102 if-then rules of the following simpler format using the same datapath: IF A and B Then Do E. With this format the chip takes two inputs and produces one outputs. We have built two VME-bus board systems based on this chip for Oak Ridge National Laboratory (ORNL). The board is now installed in a robot at ORNL. Researchers uses this board for experiment in autonomous robot navigation. The Fuzzy Logic system board places the Fuzzy chip into a VMEbus environment. High level C language functions hide the operational details of the board from the applications programme . The programmer treats rule memories and fuzzification function memories as local structures passed as parameters to the C functions. ASIC fuzzy inference hardware is extremely fast, but they are limited in generality. Many aspects of the design are limited or fixed. We have proposed to designing a are limited or fixed. We have proposed to designing a fuzzy information processor as an application specific processor using a quantitative approach. The quantitative approach was developed by RISC designers. In effect, we are interested in evaluating the effectiveness of a specialized RISC processor for fuzzy information processing. As the first step, we measured the possible speed-up of a fuzzy inference program based on if-then rules by an introduction of specialized instructions, i.e., min and max instructions. The minimum and maximum operations are heavily used in fuzzy logic applications as fuzzy intersection and union. We performed measurements using a MIPS R3000 as a base micropro essor. The initial result is encouraging. We can achieve as high as a 2.5 increase in inference speed if the R3000 had min and max instructions. Also, they are useful for speeding up other fuzzy operations such as bounded product and bounded sum. The embedded processor's main task is to control some device or process. It usually runs a single or a embedded processer to create an embedded processor for fuzzy control is very effective. Table I shows the measured speed of the inference by a MIPS R3000 microprocessor, a fictitious MIPS R3000 microprocessor with min and max instructions, and a UNC/MCNC ASIC fuzzy inference chip. The software that used on microprocessors is a simulator of the ASIC chip. The first row is the computation time in seconds of 6000 inferences using 51 rules where each fuzzy set is represented by an array of 64 elements. The second row is the time required to perform a single inference. The last row is the fuzzy logical inferences per second (FLIPS) measured for ach device. There is a large gap in run time between the ASIC and software approaches even if we resort to a specialized fuzzy microprocessor. As for design time and cost, these two approaches represent two extremes. An ASIC approach is extremely expensive. It is, therefore, an important research topic to design a specialized computing architecture for fuzzy applications that falls between these two extremes both in run time and design time/cost. TABLEI INFERENCE TIME BY 51 RULES {{{{Time }}{{MIPS R3000 }}{{ASIC }}{{Regular }}{{With min/mix }}{{6000 inference 1 inference FLIPS }}{{125s 20.8ms 48 }}{{49s 8.2ms 122 }}{{0.0038s 6.4㎲ 156,250 }} }}

  • PDF