• Title/Summary/Keyword: Computation reduction

Search Result 443, Processing Time 0.024 seconds

Numerical Study on the Baffle Structure for Determining the Flow Characteristic in Small Scale SCR System (소형 SCR 시스템 내 유동 제어를 위한 Baffle의 구조 결정에 관한 수치해석적 연구)

  • Park, Mi-Jung;Chang, Hyuk-Sang;Ha, Ji-Soo
    • Journal of Korean Society of Environmental Engineers
    • /
    • v.32 no.9
    • /
    • pp.862-869
    • /
    • 2010
  • Numerical analysis was done to evaluate the gas flow distribution in small scale SCR system which has $2.4{\times}2.4{\times}3.1\;m^3$ in volume and 25,300 Sm3/hr in flue gas flow capacity. Various types of baffles proposed for controlling the flow uniformity were evaluated by the CFD analysis to find the optimal geometry of the baffle in the SCR system. By installing baffles in the SCR system, the RMS (%) value was raised up to 6.2% compared with the baffle-uninstalled state. The effect of baffle thicknesses on the RMS (%) value was not shown within 0 and 8 mm in thickness, but the RMS (%) value was raised by 2.5% in 10 mm of baffles thickness, which causes the unstability in flow. By comparison between the shape of baffles, it is known that the lattice type baffle has better performance in controlling the flow uniformity than the circular truncated cone type baffle or mixer type baffle. RMS (%) values have more that 10% difference according to the shape of baffle type.

LLR-based Cooperative ARQ Protocol in Rayleigh Fading Channel (레일리 페이딩 채널에서 LLR 기반의 협력 ARQ 프로토콜)

  • Choi, Dae-Kyu;Kong, Hyung-Yun
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.45 no.4
    • /
    • pp.31-37
    • /
    • 2008
  • Conventional cooperative communications can attain gain of spatial diversity and path loss reduction because destination node independently received same signal from source node and relay node located between source node and destination node. However, these techniques bring about decreased spectral efficiency with relay node and increased complexity of receiver by using maximal ratio combining (MRC). This paper has proposed cooperative ARQ protocol that can improve the above problems and can get the better performance. This method can increase the spectral efficiency than conventional cooperative communication because if the received signal from source node is satisfied by the destination preferentially, the destination transmits ACK message to both relay node and source node and then recovers the received signal. In addition, if ARQ message indicates NACK relay node operates selective retransmission and we can increase reliability of system compared with that of general ARQ protocol in which source node retransmits data. In the proposed protocol, the selective retransmission and ARQ message are to be determined by comparing log-likelihood ratio (LLR) computation of received signal from source node with predetermined threshold values. Therefore, this protocol don't waste redundant bandwidth with CRC code and can reduce complexity of receiver without MRC. We verified spectral efficiency and BER performance for the proposed protocol through Monte-Carlo simulation over Rayleigh fading plus AWGN.

Hardware Design of High-Performance SAO in HEVC Encoder for Ultra HD Video Processing in Real Time (UHD 영상의 실시간 처리를 위한 고성능 HEVC SAO 부호화기 하드웨어 설계)

  • Cho, Hyun-pyo;Park, Seung-yong;Ryoo, Kwang-ki
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2014.10a
    • /
    • pp.271-274
    • /
    • 2014
  • This paper proposes high-performance SAO(Sample Adaptive Offset) in HEVC(High Efficiency Video Coding) encoder for Ultra HD video processing in real time. SAO is a newly adopted technique belonging to the in-loop filter in HEVC. The proposed SAO encoder hardware architecture uses three-layered buffers to minimize memory access time and to simplify pixel processing and also uses only adder, subtractor, shift register and feed-back comparator to reduce area. Furthermore, the proposed architecture consists of pipelined pixel classification and applying SAO parameters, and also classifies four consecutive pixels into EO and BO concurrently. These result in the reduction of processing time and computation. The proposed SAO encoder architecture is designed by Verilog HDL, and implemented by 180k logic gates in TSMC $0.18{\mu}m$ process. At 110MHz, the proposed SAO encoder can support 4K Ultra HD video encoding at 30fps in real time.

  • PDF

Vital Sign Detection in a Noisy Environment by Undesirable Micro-Motion (원하지 않는 작은 동작에 의한 잡음 환경 내 생체신호 탐지 기법)

  • Choi, In-Oh;Kim, Min;Choi, Jea-Ho;Park, Jeong-Ki;Kim, Kyung-Tae
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.30 no.5
    • /
    • pp.418-426
    • /
    • 2019
  • Recently, many studies on vital sign detection using a radar sensor related to Internet of Things(IoT) smart home systems have been conducted. Because vital signs such as respiration and cardiac rates generally cause micro-motions in the chest or back, the phase of the received echo signal from a target fluctuates according to the micro-motion. Therefore, vital signs are usually detected via spectral analysis of the phase. However, the probability of false alarms in cardiac rate detection increases as a result of various problems in the measurement environment, such as very weak phase fluctuations caused by the cardiac rate. Therefore, this study analyzes the difficulties of vital sign detection and proposes an efficient vital sign detection algorithm consisting of four main stages: 1) phase decomposition, 2) phase differentiation and filtering, 3) vital sign detection, and 4) reduction of the probability of false alarm. Experimental results using impulse-radio ultra-wideband radar show that the proposed algorithm is very efficient in terms of computation and accuracy.

Active VM Consolidation for Cloud Data Centers under Energy Saving Approach

  • Saxena, Shailesh;Khan, Mohammad Zubair;Singh, Ravendra;Noorwali, Abdulfattah
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.11
    • /
    • pp.345-353
    • /
    • 2021
  • Cloud computing represent a new era of computing that's forms through the combination of service-oriented architecture (SOA), Internet and grid computing with virtualization technology. Virtualization is a concept through which every cloud is enable to provide on-demand services to the users. Most IT service provider adopt cloud based services for their users to meet the high demand of computation, as it is most flexible, reliable and scalable technology. Energy based performance tradeoff become the main challenge in cloud computing, as its acceptance and popularity increases day by day. Cloud data centers required a huge amount of power supply to the virtualization of servers for maintain on- demand high computing. High power demand increase the energy cost of service providers as well as it also harm the environment through the emission of CO2. An optimization of cloud computing based on energy-performance tradeoff is required to obtain the balance between energy saving and QoS (quality of services) policies of cloud. A study about power usage of resources in cloud data centers based on workload assign to them, says that an idle server consume near about 50% of its peak utilization power [1]. Therefore, more number of underutilized servers in any cloud data center is responsible to reduce the energy performance tradeoff. To handle this issue, a lots of research proposed as energy efficient algorithms for minimize the consumption of energy and also maintain the SLA (service level agreement) at a satisfactory level. VM (virtual machine) consolidation is one such technique that ensured about the balance of energy based SLA. In the scope of this paper, we explore reinforcement with fuzzy logic (RFL) for VM consolidation to achieve energy based SLA. In this proposed RFL based active VM consolidation, the primary objective is to manage physical server (PS) nodes in order to avoid over-utilized and under-utilized, and to optimize the placement of VMs. A dynamic threshold (based on RFL) is proposed for over-utilized PS detection. For over-utilized PS, a VM selection policy based on fuzzy logic is proposed, which selects VM for migration to maintain the balance of SLA. Additionally, it incorporate VM placement policy through categorization of non-overutilized servers as- balanced, under-utilized and critical. CloudSim toolkit is used to simulate the proposed work on real-world work load traces of CoMon Project define by PlanetLab. Simulation results shows that the proposed policies is most energy efficient compared to others in terms of reduction in both electricity usage and SLA violation.

Development of Numerical Computation Techniques for the Free-Surface of U-Tube Type Anti-roll Tank (U-튜브형 횡동요 감쇄 탱크의 자유수면 해석기법 개발에 관한 연구)

  • Sang-Eui Lee
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.28 no.7
    • /
    • pp.1244-1251
    • /
    • 2022
  • Marine accidents due to a loss of stability, have been gradually increasing over the last decade. Measures must be taken on the roll reduction of a ship. Amongst the measures, building an anti-roll tank in a ship is recognized as the most simple and effective way to reduce the roll motion. Therefore, this study aims to develop a computational model for a U-tube type anti-roll tank and to validate it by experiment. In particular, to validate the developed computational model, the height of the free surface in the tank was measured in the experiment. To develop a computational model, the mesh dependency test was carried out. Further, the effects of a turbulence model, time step size, and the number of iterations on the numerical solution were analyzed. In summary, a U-tube type anti-roll tank simulation had to be performed accurately with conditions of a realizable k-𝜖 turbulence model, 10-2s time step size, and 15 iterations. In validation, the two cases of measured data from the experiment were compared with the numerical results. In the present study, STAR-CCM+ (ver. 17.02), a RANS-based commercial solver was used.

A Study about Learning Graph Representation on Farmhouse Apple Quality Images with Graph Transformer (그래프 트랜스포머 기반 농가 사과 품질 이미지의 그래프 표현 학습 연구)

  • Ji Hun Bae;Ju Hwan Lee;Gwang Hyun Yu;Gyeong Ju Kwon;Jin Young Kim
    • Smart Media Journal
    • /
    • v.12 no.1
    • /
    • pp.9-16
    • /
    • 2023
  • Recently, a convolutional neural network (CNN) based system is being developed to overcome the limitations of human resources in the apple quality classification of farmhouse. However, since convolutional neural networks receive only images of the same size, preprocessing such as sampling may be required, and in the case of oversampling, information loss of the original image such as image quality degradation and blurring occurs. In this paper, in order to minimize the above problem, to generate a image patch based graph of an original image and propose a random walk-based positional encoding method to apply the graph transformer model. The above method continuously learns the position embedding information of patches which don't have a positional information based on the random walk algorithm, and finds the optimal graph structure by aggregating useful node information through the self-attention technique of graph transformer model. Therefore, it is robust and shows good performance even in a new graph structure of random node order and an arbitrary graph structure according to the location of an object in an image. As a result, when experimented with 5 apple quality datasets, the learning accuracy was higher than other GNN models by a minimum of 1.3% to a maximum of 4.7%, and the number of parameters was 3.59M, which was about 15% less than the 23.52M of the ResNet18 model. Therefore, it shows fast reasoning speed according to the reduction of the amount of computation and proves the effect.

Improved Social Network Analysis Method in SNS (SNS에서의 개선된 소셜 네트워크 분석 방법)

  • Sohn, Jong-Soo;Cho, Soo-Whan;Kwon, Kyung-Lag;Chung, In-Jeong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.117-127
    • /
    • 2012
  • Due to the recent expansion of the Web 2.0 -based services, along with the widespread of smartphones, online social network services are being popularized among users. Online social network services are the online community services which enable users to communicate each other, share information and expand human relationships. In the social network services, each relation between users is represented by a graph consisting of nodes and links. As the users of online social network services are increasing rapidly, the SNS are actively utilized in enterprise marketing, analysis of social phenomenon and so on. Social Network Analysis (SNA) is the systematic way to analyze social relationships among the members of the social network using the network theory. In general social network theory consists of nodes and arcs, and it is often depicted in a social network diagram. In a social network diagram, nodes represent individual actors within the network and arcs represent relationships between the nodes. With SNA, we can measure relationships among the people such as degree of intimacy, intensity of connection and classification of the groups. Ever since Social Networking Services (SNS) have drawn increasing attention from millions of users, numerous researches have made to analyze their user relationships and messages. There are typical representative SNA methods: degree centrality, betweenness centrality and closeness centrality. In the degree of centrality analysis, the shortest path between nodes is not considered. However, it is used as a crucial factor in betweenness centrality, closeness centrality and other SNA methods. In previous researches in SNA, the computation time was not too expensive since the size of social network was small. Unfortunately, most SNA methods require significant time to process relevant data, and it makes difficult to apply the ever increasing SNS data in social network studies. For instance, if the number of nodes in online social network is n, the maximum number of link in social network is n(n-1)/2. It means that it is too expensive to analyze the social network, for example, if the number of nodes is 10,000 the number of links is 49,995,000. Therefore, we propose a heuristic-based method for finding the shortest path among users in the SNS user graph. Through the shortest path finding method, we will show how efficient our proposed approach may be by conducting betweenness centrality analysis and closeness centrality analysis, both of which are widely used in social network studies. Moreover, we devised an enhanced method with addition of best-first-search method and preprocessing step for the reduction of computation time and rapid search of the shortest paths in a huge size of online social network. Best-first-search method finds the shortest path heuristically, which generalizes human experiences. As large number of links is shared by only a few nodes in online social networks, most nods have relatively few connections. As a result, a node with multiple connections functions as a hub node. When searching for a particular node, looking for users with numerous links instead of searching all users indiscriminately has a better chance of finding the desired node more quickly. In this paper, we employ the degree of user node vn as heuristic evaluation function in a graph G = (N, E), where N is a set of vertices, and E is a set of links between two different nodes. As the heuristic evaluation function is used, the worst case could happen when the target node is situated in the bottom of skewed tree. In order to remove such a target node, the preprocessing step is conducted. Next, we find the shortest path between two nodes in social network efficiently and then analyze the social network. For the verification of the proposed method, we crawled 160,000 people from online and then constructed social network. Then we compared with previous methods, which are best-first-search and breath-first-search, in time for searching and analyzing. The suggested method takes 240 seconds to search nodes where breath-first-search based method takes 1,781 seconds (7.4 times faster). Moreover, for social network analysis, the suggested method is 6.8 times and 1.8 times faster than betweenness centrality analysis and closeness centrality analysis, respectively. The proposed method in this paper shows the possibility to analyze a large size of social network with the better performance in time. As a result, our method would improve the efficiency of social network analysis, making it particularly useful in studying social trends or phenomena.

Ensemble Learning with Support Vector Machines for Bond Rating (회사채 신용등급 예측을 위한 SVM 앙상블학습)

  • Kim, Myoung-Jong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.29-45
    • /
    • 2012
  • Bond rating is regarded as an important event for measuring financial risk of companies and for determining the investment returns of investors. As a result, it has been a popular research topic for researchers to predict companies' credit ratings by applying statistical and machine learning techniques. The statistical techniques, including multiple regression, multiple discriminant analysis (MDA), logistic models (LOGIT), and probit analysis, have been traditionally used in bond rating. However, one major drawback is that it should be based on strict assumptions. Such strict assumptions include linearity, normality, independence among predictor variables and pre-existing functional forms relating the criterion variablesand the predictor variables. Those strict assumptions of traditional statistics have limited their application to the real world. Machine learning techniques also used in bond rating prediction models include decision trees (DT), neural networks (NN), and Support Vector Machine (SVM). Especially, SVM is recognized as a new and promising classification and regression analysis method. SVM learns a separating hyperplane that can maximize the margin between two categories. SVM is simple enough to be analyzed mathematical, and leads to high performance in practical applications. SVM implements the structuralrisk minimization principle and searches to minimize an upper bound of the generalization error. In addition, the solution of SVM may be a global optimum and thus, overfitting is unlikely to occur with SVM. In addition, SVM does not require too many data sample for training since it builds prediction models by only using some representative sample near the boundaries called support vectors. A number of experimental researches have indicated that SVM has been successfully applied in a variety of pattern recognition fields. However, there are three major drawbacks that can be potential causes for degrading SVM's performance. First, SVM is originally proposed for solving binary-class classification problems. Methods for combining SVMs for multi-class classification such as One-Against-One, One-Against-All have been proposed, but they do not improve the performance in multi-class classification problem as much as SVM for binary-class classification. Second, approximation algorithms (e.g. decomposition methods, sequential minimal optimization algorithm) could be used for effective multi-class computation to reduce computation time, but it could deteriorate classification performance. Third, the difficulty in multi-class prediction problems is in data imbalance problem that can occur when the number of instances in one class greatly outnumbers the number of instances in the other class. Such data sets often cause a default classifier to be built due to skewed boundary and thus the reduction in the classification accuracy of such a classifier. SVM ensemble learning is one of machine learning methods to cope with the above drawbacks. Ensemble learning is a method for improving the performance of classification and prediction algorithms. AdaBoost is one of the widely used ensemble learning techniques. It constructs a composite classifier by sequentially training classifiers while increasing weight on the misclassified observations through iterations. The observations that are incorrectly predicted by previous classifiers are chosen more often than examples that are correctly predicted. Thus Boosting attempts to produce new classifiers that are better able to predict examples for which the current ensemble's performance is poor. In this way, it can reinforce the training of the misclassified observations of the minority class. This paper proposes a multiclass Geometric Mean-based Boosting (MGM-Boost) to resolve multiclass prediction problem. Since MGM-Boost introduces the notion of geometric mean into AdaBoost, it can perform learning process considering the geometric mean-based accuracy and errors of multiclass. This study applies MGM-Boost to the real-world bond rating case for Korean companies to examine the feasibility of MGM-Boost. 10-fold cross validations for threetimes with different random seeds are performed in order to ensure that the comparison among three different classifiers does not happen by chance. For each of 10-fold cross validation, the entire data set is first partitioned into tenequal-sized sets, and then each set is in turn used as the test set while the classifier trains on the other nine sets. That is, cross-validated folds have been tested independently of each algorithm. Through these steps, we have obtained the results for classifiers on each of the 30 experiments. In the comparison of arithmetic mean-based prediction accuracy between individual classifiers, MGM-Boost (52.95%) shows higher prediction accuracy than both AdaBoost (51.69%) and SVM (49.47%). MGM-Boost (28.12%) also shows the higher prediction accuracy than AdaBoost (24.65%) and SVM (15.42%)in terms of geometric mean-based prediction accuracy. T-test is used to examine whether the performance of each classifiers for 30 folds is significantly different. The results indicate that performance of MGM-Boost is significantly different from AdaBoost and SVM classifiers at 1% level. These results mean that MGM-Boost can provide robust and stable solutions to multi-classproblems such as bond rating.

Investigation on Characteristics of Swine Manure of Optimum Volume for Escalator Reversing Composting Facility (돼지분뇨 특성에 따른 기계교반 퇴비화시설의 적정용적 산정 연구)

  • Kwag, J.H.;Choi, D.Y.;Park, C.H.;Jeong, K.H.;Kim, J.H.;Yoo, Y.H.;Youn, C.K.;Ra, C.S.
    • Journal of Animal Environmental Science
    • /
    • v.14 no.2
    • /
    • pp.105-112
    • /
    • 2008
  • This study was carried out to investigate evaporation rate of moisture per surface area and degradation rate of organic matter in full scale escalator reversing composting facility were analyzed to develope a computer program for the computation of an optimum volume of composting facility according to handling methods of swine farm, moisture levels of manure, degradation rate of organics and evaporation rate of moisture during composting. The obtained results can be followed as bellow; The temperature in full scale escalator reversing composting facility during composting reached $70^{\circ}C$ in 4 days and maintained until 11 days. Reduction rate of moisture and density was average 1.20% and 29.7%, respectively. Annual degradation rate of organic matter was 3.53%, showing lowest rate in winter as 3.23%. These seasonal degradation rate could be a factor to be considered for proper management and installation of composting facility. When computed with the amount of feces, urine, slurry and manure plus wastewater produced, the optimum volumes of composting facility for slurry and manure plus wastewater including each 95% moisture was $229m^3$ and $277m^3$, respectively, showing 21% ($48m^3$) difference.

  • PDF