• Title/Summary/Keyword: Internet Round

Search Result 162, Processing Time 0.024 seconds

Two-round ID-based Group Key Agreement Fitted for Pay-TV System (유료 방송 시스템에 적합한 ID기반의 2 라운드 그룹키 동의 프로토콜)

  • Kim Hyunjue;Nam Junghyun;Kim Seungjoo;Won Dongho
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.15 no.1
    • /
    • pp.41-55
    • /
    • 2005
  • A group key agreement protocol allows a group of user to share a key which may later be used to achieve certain cryptographic goals. In this paper, we propose a new scalable two-round ID-based group key agreement protocol which would be well fit to a Pay-TV system, additionally. to the fields of internet stock quotes, audio and music deliveries, software updates and the like. Our protocol improves the three round poop key agreement protocol of Nam et al., resulting in upgrading the computational efficiency by using the batch verification technique in pairing-based cryptography. Also our protocol simplifies the key agreement procedures by utilizing ID-based system. We prove the security of our protocol under the Computational Diffie-Hellman assumption and the Bilinear Decisional Diffie-Hellman assumption. Also we analyze its efficiency.

Estimation of TCP Throughput Fairness Ratio under Various Background Traffic (다양한 백그라운드 트래픽이 존재하는 경우의 TCP 공정성 비율 측정)

  • Lee, Jun-Soo;Kim, Ju-Kyun
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.2
    • /
    • pp.197-205
    • /
    • 2008
  • TCP packets occupy over 90% of current Internet traffic thus understanding of TCP throughput is crucial to understand Internet. Under the TCP congestion regime, heterogeneous flows, i.e., flows with different round-trip times (RTTs), that share the same bottleneck link will not attain equal portions of the available bandwidth. In fact, according to the TCP friendly formula, the throughput ratio of two flows is inversely proportional to the ratio of their RTTs. It has also been shown that TCP's unfairness to flows with longer RTTs is accentuated under loss synchronization. In this paper, we show that, injecting bursty background traffic may actually lead to new type of synchronization and result in unfairness to foreground TCP flows with longer RTTs. We propose three different metrics to characterize traffic burstiness and show that these metrics are reliable predictors of TCP unfairness.

  • PDF

Analytical Modelling and Heuristic Algorithm for Object Transfer Latency in the Internet of Things (사물인터넷에서 객체전송지연을 계산하기 위한 수리적 모델링 및 휴리스틱 알고리즘의 개발)

  • Lee, Yong-Jin
    • Journal of Internet of Things and Convergence
    • /
    • v.6 no.3
    • /
    • pp.1-6
    • /
    • 2020
  • This paper aims to integrate the previous models about mean object transfer latency in one framework and analyze the result through the computational experience. The analytical object transfer latency model assumes the multiple packet losses and the Internet of Things(IoT) environment including multi-hop wireless network, where fast re-transmission is not possible due to small window. The model also considers the initial congestion window size and the multiple packet loss in one congestion window. Performance evaluation shows that the lower and upper bounds of the mean object transfer latency are almost the same when both transfer object size and packet loss rate are small. However, as packet loss rate increases, the size of the initial congestion window and the round-trip time affect the upper and lower bounds of the mean object transfer latency.

Evaluation of MOS for the Access Delay of Internet Service (인터넷 서비스의 접속지연에 대한 MOS 평가)

  • Lee, Hoon;Lee, Young-Ok
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.9B
    • /
    • pp.900-907
    • /
    • 2009
  • With the introduction of the concept such as QoS, QoE, and MOS for the multimedia services over the high-speed network the user's concern about the degradation in the perceived quality of service for the conventional data service has become a new problem to the Internet service providers. In this work we propose a method to evaluate the subjective QoE such as MOS for the high-speed Internet service, where the round trip time for the data transfer is taken into account as a performance parameter for the experienced delay. We develop a tool to evaluate an MOS that enables us to investigate the behavior of users for diverse delay times for the web access service, via which we analyze the sensitivity of users with regard to MOS as a function of delay.

Analytical model for mean web object transfer latency estimation in the narrowband IoT environment (협대역 사물 인터넷 환경에서 웹 객체의 평균 전송시간을 추정하기 위한 해석적 모델)

  • Lee, Yong-Jin
    • Journal of Internet of Things and Convergence
    • /
    • v.1 no.1
    • /
    • pp.1-4
    • /
    • 2015
  • This paper aims to present the mathematical model to find the mean web object transfer latency in the slow-start phase of TCP congestion control mechanism, which is one of the main control techniques of Internet. Mean latency is an important service quality measure of end-user in the network. The application area of the proposed latency model is the narrowband environment including multi-hop wireless network and Internet of Things(IoT), where packet loss occurs in the slow-start phase only due to small window. The model finds the latency considering initial window size and the packet loss rate. Our model shows that for a given packet loss rate, round trip time and initial window size mainly affect the mean web object transfer latency. The proposed model can be applied to estimate the mean response time that end user requires in the IoT service applications.

Implementation of Hair Style Recommendation System Based on Big data and Deepfakes (빅데이터와 딥페이크 기반의 헤어스타일 추천 시스템 구현)

  • Tae-Kook Kim
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.3
    • /
    • pp.13-19
    • /
    • 2023
  • In this paper, we investigated the implementation of a hairstyle recommendation system based on big data and deepfake technology. The proposed hairstyle recommendation system recognizes the facial shapes based on the user's photo (image). Facial shapes are classified into oval, round, and square shapes, and hairstyles that suit each facial shape are synthesized using deepfake technology and provided as videos. Hairstyles are recommended based on big data by applying the latest trends and styles that suit the facial shape. With the image segmentation map and the Motion Supervised Co-Part Segmentation algorithm, it is possible to synthesize elements between images belonging to the same category (such as hair, face, etc.). Next, the synthesized image with the hairstyle and a pre-defined video are applied to the Motion Representations for Articulated Animation algorithm to generate a video animation. The proposed system is expected to be used in various aspects of the beauty industry, including virtual fitting and other related areas. In future research, we plan to study the development of a smart mirror that recommends hairstyles and incorporates features such as Internet of Things (IoT) functionality.

A Congestion Control Algorithm for the fairness Improvement of TCP Vegas (TCP Vegas의 공정성 향상을 위한 혼잡 제어 알고리즘)

  • 오민철;송병훈;정광수
    • Journal of KIISE:Information Networking
    • /
    • v.31 no.3
    • /
    • pp.269-279
    • /
    • 2004
  • The most important factor influencing the robustness of the Internet Is the end-to-end TCP congestion control. However, the congestion control scheme of TCP Reno, the most popular TCP version on the Internet, employs passive congestion indication. It makes worse the network congestion. Recently, Brakmo and Peterson have proposed a new version of TCP, which is named TCP Vegas, with a fundamentally different congestion control scheme from that of the Reno. Many studies indicate that the Vegas is able to achieve better throughput and higher stability than the Reno. But there are two unfairness problems in Vegas. These problems hinder the spread of the Vegas in current Internet. In this paper, in order to solve these unfairness problems, we propose a new congestion control algorithm called TCP PowerVegas. The existing Vegas depends mainly only on the rtt(round trip time), but the proposed PowerVegas use the new congestion control scheme combined the Information on the rtt with the information on the packet loss. Therefore the PowerVegas performs the congestion control more competitively than the Vegas. Thus, the PowerVegas is able to solve effectively these unfairness problems which the Vegas has experienced. To evaluate the proposed approach, we compare the performance among PowerVegas, Reno and Vegas under same network environment. Using simulation, the PowerVegas is able to achieve better throughput and higher stability than the Reno and is shown to achieve much better fairness than the existing Vegas.

Neighborhood coreness algorithm for identifying a set of influential spreaders in complex networks

  • YANG, Xiong;HUANG, De-Cai;ZHANG, Zi-Ke
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.6
    • /
    • pp.2979-2995
    • /
    • 2017
  • In recent years, there has been an increasing number of studies focused on identifying a set of spreaders to maximize the influence of spreading in complex networks. Although the k-core decomposition can effectively identify the single most influential spreader, selecting a group of nodes that has the largest k-core value as the seeds cannot increase the performance of the influence maximization because the propagation sphere of this group of nodes is overlapped. To overcome this limitation, we propose a neighborhood coreness cover and discount heuristic algorithm named "NCCDH" to identify a set of influential and decentralized seeds. Using this method, a node in the high-order shell with the largest neighborhood coreness and an uncovered status will be selected as the seed in each turn. In addition, the neighbors within the same shell layer of this seed will be covered, and the neighborhood coreness of the neighbors outside the shell layer will be discounted in the subsequent round. The experimental results show that with increases in the spreading probability, the NCCDH outperforms other algorithms in terms of the affected scale and spreading speed under the Susceptible-Infected-Recovered (SIR) and Susceptible-Infected (SI) models. Furthermore, this approach has a superior running time.

Real Time Related Key Attack on Hummingbird-2

  • Zhang, Kai;Ding, Lin;Li, Junzhi;Guan, Jie
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.8
    • /
    • pp.1946-1963
    • /
    • 2012
  • Hummingbird is a lightweight encryption and message authentication primitive published in RISC'09 and WLC'10. In FSE'11, Markku-Juhani O.Saarinen presented a differential divide-and-conquer method which has complexity upper bounded by $2^{64}$ operations and requires processing of few megabytes of chosen messages under two related nonces (IVs). The improved version, Hummingbird-2, was presented in RFIDSec 2011. Based on the idea of differential collision, this paper discovers some weaknesses of the round function WD16. Combining with the simple key loading algorithm, a related-key chosen-IV attack which can recover the full secret key is proposed. Under 15 pairs of related keys, the 128 bit initial key can be recovered, requiring $2^{27}$ chosen IV and the computational complexity is $O(2^{27})$. In average, the attack needs several minutes to recover the full 128-bit secret key on a PC. The experimental result corroborates our attack. The result shows that the Hummingbird-2 cipher can't resist related key attack.

An Improved Privacy Preserving Construction for Data Integrity Verification in Cloud Storage

  • Xia, Yingjie;Xia, Fubiao;Liu, Xuejiao;Sun, Xin;Liu, Yuncai;Ge, Yi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.10
    • /
    • pp.3607-3623
    • /
    • 2014
  • The increasing demand in promoting cloud computing in either business or other areas requires more security of a cloud storage system. Traditional cloud storage systems fail to protect data integrity information (DII), when the interactive messages between the client and the data storage server are sniffed. To protect DII and support public verifiability, we propose a data integrity verification scheme by deploying a designated confirmer signature DCS as a building block. The DCS scheme strikes the balance between public verifiable signatures and zero-knowledge proofs which can address disputes between the cloud storage server and any user, whoever acting as a malicious player during the two-round verification. In addition, our verification scheme remains blockless and stateless, which is important in conducting a secure and efficient cryptosystem. We perform security analysis and performance evaluation on our scheme, and compared with the existing schemes, the results show that our scheme is more secure and efficient.