• Title/Summary/Keyword: Vector Image

Search Result 1,574, Processing Time 0.024 seconds

A Study on the Improvement of UAV based 3D Point Cloud Spatial Object Location Accuracy using Road Information (도로정보를 활용한 UAV 기반 3D 포인트 클라우드 공간객체의 위치정확도 향상 방안)

  • Lee, Jaehee;Kang, Jihun;Lee, Sewon
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.5_1
    • /
    • pp.705-714
    • /
    • 2019
  • Precision positioning is necessary for various use of high-resolution UAV images. Basically, GCP is used for this purpose, but in case of emergency situations or difficulty in selecting GCPs, the data shall be obtained without GCPs. This study proposed a method of improving positional accuracy for x, y coordinate of UAV based 3 dimensional point cloud data generated without GCPs. Road vector file by the public data (Open Data Portal) was used as reference data for improving location accuracy. The geometric correction of the 2 dimensional ortho-mosaic image was first performed and the transform matrix produced in this process was adopted to apply to the 3 dimensional point cloud data. The straight distance difference of 34.54 m before the correction was reduced to 1.21 m after the correction. By confirming that it is possible to improve the location accuracy of UAV images acquired without GCPs, it is expected to expand the scope of use of 3 dimensional spatial objects generated from point cloud by enabling connection and compatibility with other spatial information data.

Hand Motion Recognition Algorithm Using Skin Color and Center of Gravity Profile (피부색과 무게중심 프로필을 이용한 손동작 인식 알고리즘)

  • Park, Youngmin
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.2
    • /
    • pp.411-417
    • /
    • 2021
  • The field that studies human-computer interaction is called HCI (Human-computer interaction). This field is an academic field that studies how humans and computers communicate with each other and recognize information. This study is a study on hand gesture recognition for human interaction. This study examines the problems of existing recognition methods and proposes an algorithm to improve the recognition rate. The hand region is extracted based on skin color information for the image containing the shape of the human hand, and the center of gravity profile is calculated using principal component analysis. I proposed a method to increase the recognition rate of hand gestures by comparing the obtained information with predefined shapes. We proposed a method to increase the recognition rate of hand gestures by comparing the obtained information with predefined shapes. The existing center of gravity profile has shown the result of incorrect hand gesture recognition for the deformation of the hand due to rotation, but in this study, the center of gravity profile is used and the point where the distance between the points of all contours and the center of gravity is the longest is the starting point. Thus, a robust algorithm was proposed by re-improving the center of gravity profile. No gloves or special markers attached to the sensor are used for hand gesture recognition, and a separate blue screen is not installed. For this result, find the feature vector at the nearest distance to solve the misrecognition, and obtain an appropriate threshold to distinguish between success and failure.

Analysis of Mashup Performances based on Vector Layer of Various GeoWeb 2.0 Platform Open APIs (다양한 공간정보 웹 2.0 플랫폼 Open API의 벡터 레이어 기반 매쉬업 성능 분석)

  • Kang, Jinwon;Kim, Min-soo
    • Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology
    • /
    • v.9 no.4
    • /
    • pp.745-754
    • /
    • 2019
  • As GeoWeb 2.0 technologies are widely used, various kinds of services that mashup spatial data and user data are being developed. In particular, various spatial information platforms such as Google Maps, OpenStreetMap, Daum Map, Naver Map, olleh Map, and VWorld based on GeoWeb 2.0 technologies support mashup service. The mashup service which is supported through the Open APIs of the platforms, provides various kinds of spatial data such as 2D map, 3D map, and aerial image. Also, application fields using the mashup service are greatly expanded. Recently, as user data for mashup have been greatly increased, there was a problem in mashup performance. However, the research on the mashup performance improvement is currently insufficient, even the research on the mashup performance comparison of various platforms has not been performed. In this paper, we perform comparative analysis of the mashup performance for large amounts of user data and spatial data using various spatial information platforms available in Korea. Specifically, we propose two performance analysis indexes of mashup time and user interaction time in order to analyze the mashup performance efficiently. Also, we implement a system for the performance analysis. Finally, from the performance analysis result, we propose a spatial information platform that can be efficiently applied to cases when user data increases greatly and user interaction occurs frequently.

Accuracy Evaluation of Pre- and Post-treatment Setup Errors in CBCT-based Stereotactic Body Radiation Therapy (SBRT) for Lung Tumor (CBCT 기반 폐 종양 정위 신체 방사선 요법(SBRT)에서 치료 전·후 set up 에러의 정확도 평가)

  • Jang, Eun-Sung;Choi, Ji-Hoon
    • Journal of the Korean Society of Radiology
    • /
    • v.15 no.6
    • /
    • pp.861-867
    • /
    • 2021
  • Since SBRT takes up to 1 hour from 30 minutes to treatment fraction once or three to five times, there is a possibility of setup error during treatment. To reduce these set-up errors and give accurate doses, we intend to evaluate the usefulness of pre-treatment and post-treatment error values by imaging CBCT again to determine postural movement due to pre-treatment coordinate values using pre-treatment CBCT. On average, the range of systematic errors was 0.032 to 0.17 on the X and Y,Z axes, confirming that there was very little change in movement even after treatment. Tumor centripetal changes (±SD) due to respiratory tuning were 0.11 (±0.12) cm, 0.27 (±0.15) cm, and 0.21 cm (±0.31 cm) in the X, Y and Z directions. The tumor edges ±SD were 0.21 (±0.18) cm, 0.30 (±0.23) cm, and 0.19 cm (±0.26) cm in the X, Y and Z directions. The (±SD) of tumor-corrected displacements were 0.03 (±0.16) cm, 0.05 (±0.26) cm, and 0.02 (±0.23) cm in RL, AP, and SI directions, respectively. The range of the 3D vector value was 0.11 to 0-.18 cm on average when comparing pre-treatment and CBCT, and it was confirmed that the corrected set-up error was within 0.3 cm. Therefore, it was confirmed that there were some changes in values depending on some older patients, condition on the day of treatment, and body type, but they were within the significance range.

A Study on the Length of DMZ and MDL (비무장지대 및 군사분계선의 길이에 관한 연구)

  • KIM, Chang-Hwan
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.22 no.1
    • /
    • pp.19-27
    • /
    • 2019
  • This study is to measure the length of the Demilitarized Zone and the Military Demarcation Line(MDL) on the Korean Peninsular. For this purpose, maps of the Armistice Agreement Volume II were used. These maps are nine sheets. In order to extract the MDL shown on the map, coordinates were assigned to the scanned image maps using the georeferencing module of ArcGIS based on the sheet line coordinates. The accuracy of the extracted vectors was checked by overlaying them on the maps of the Armistice Agreement Volume II. And I tried to validate these vectors through comparative analysis with vectors extracted from Kim(2007). Vectors extracted from Kim(2007) had errors in the curvilinear parts of the MDL, but the vectors extracted from this study exactly matched the MDL in the Armistice Agreement Volume II. The measured length is 239.42km(148.77miles). This means that the expression '155mile MDL' or '248km DMZ' in papers, reports or mass media has so far been inappropriate. I think this study will be able to provide information on the exact length of the DMZ in studies related with DMZ or in policy decisions by the national and local government. However, it is deemed necessary to verify this result by national organizations such as the NGII(National Geographic Information Institute). After these verification procedures, I hope that the national government will inform the people of the exact length of DMZ and MDL.

Reproducibility evaluation of the use of pressure conserving abdominal compressor in lung and liver volumetric modulated arc therapy (흉복부 방사선 치료 시 압력 기반 복부압박장치 적용에 따른 치료 간 재현성 평가)

  • Park, ga yeon;Kim, joo ho;Shin, hyun kyung;Kim, min soo
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.33
    • /
    • pp.71-78
    • /
    • 2021
  • Purpose: To evaluate the inter-fractional position and respiratory reproducibility of lung and liver tumors using pressure conserving type(P-type) abdominal compressor in volumetric modulated arc therapy(VMAT). Materials and methods: Six lung cancer patients and three liver cancer patients who underwent VMAT using a P-type abdominal compressor were included in this study. Cone-beam computed tomography(CBCT) images were acquired before each treatment and compared with planning CT images to evaluate the inter-fractional position reproducibility. The position variation was defined as the difference of position shift values between target matching and bone matching. 4-dimensional cone-beam computed tomography(4D CBCT) images were acquired weekly before treatment and compared with planning 4DCT images to evaluate the inter-fractional respiratory reproducibility. The respiratory variation was calculated by the magnitude of excursions by breathing. Results: The mean ± standard deviation(SD) of overall position variation values, 3D vector in the three translational directions were 1.1 ± 1.4 mm and 4.5 ± 2.8 mm for the lung and liver, respectively. The mean ± SD of respiratory variation values were 0.7 ± 3.4 mm (p = 0.195) in the lung and 3.6 ± 2.6 mm (p < 0.05) in the liver. Conclusion: The use of P-type compressor in lung and liver VMAT was effective for stable control of inter-fractional position and respiratory variation by reproduction of abdominal compression. Appropriate PTV margin must be considered in treatment planning, and image guidance before each treatment are required in order to obtain more stable reproducibility

A Study on Improving Facial Recognition Performance to Introduce a New Dog Registration Method (새로운 반려견 등록방식 도입을 위한 안면 인식 성능 개선 연구)

  • Lee, Dongsu;Park, Gooman
    • Journal of Broadcast Engineering
    • /
    • v.27 no.5
    • /
    • pp.794-807
    • /
    • 2022
  • Although registration of dogs is mandatory according to the revision of the Animal Protection Act, the registration rate is low due to the inconvenience of the current registration method. In this paper, a performance improvement study was conducted on the dog face recognition technology, which is being reviewed as a new registration method. Through deep learning learning, an embedding vector for facial recognition of a dog was created and a method for identifying each dog individual was experimented. We built a dog image dataset for deep learning learning and experimented with InceptionNet and ResNet-50 as backbone networks. It was learned by the triplet loss method, and the experiments were divided into face verification and face recognition. In the ResNet-50-based model, it was possible to obtain the best facial verification performance of 93.46%, and in the face recognition test, the highest performance of 91.44% was obtained in rank-5, respectively. The experimental methods and results presented in this paper can be used in various fields, such as checking whether a dog is registered or not, and checking an object at a dog access facility.

A Vision Transformer Based Recommender System Using Side Information (부가 정보를 활용한 비전 트랜스포머 기반의 추천시스템)

  • Kwon, Yujin;Choi, Minseok;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.3
    • /
    • pp.119-137
    • /
    • 2022
  • Recent recommendation system studies apply various deep learning models to represent user and item interactions better. One of the noteworthy studies is ONCF(Outer product-based Neural Collaborative Filtering) which builds a two-dimensional interaction map via outer product and employs CNN (Convolutional Neural Networks) to learn high-order correlations from the map. However, ONCF has limitations in recommendation performance due to the problems with CNN and the absence of side information. ONCF using CNN has an inductive bias problem that causes poor performances for data with a distribution that does not appear in the training data. This paper proposes to employ a Vision Transformer (ViT) instead of the vanilla CNN used in ONCF. The reason is that ViT showed better results than state-of-the-art CNN in many image classification cases. In addition, we propose a new architecture to reflect side information that ONCF did not consider. Unlike previous studies that reflect side information in a neural network using simple input combination methods, this study uses an independent auxiliary classifier to reflect side information more effectively in the recommender system. ONCF used a single latent vector for user and item, but in this study, a channel is constructed using multiple vectors to enable the model to learn more diverse expressions and to obtain an ensemble effect. The experiments showed our deep learning model improved performance in recommendation compared to ONCF.

A Development of Automatic Lineament Extraction Algorithm from Landsat TM images for Geological Applications (지질학적 활용을 위한 Landsat TM 자료의 자동화된 선구조 추출 알고리즘의 개발)

  • 원중선;김상완;민경덕;이영훈
    • Korean Journal of Remote Sensing
    • /
    • v.14 no.2
    • /
    • pp.175-195
    • /
    • 1998
  • Automatic lineament extraction algorithms had been developed by various researches for geological purpose using remotely sensed data. However, most of them are designed for a certain topographic model, for instance rugged mountainous region or flat basin. Most of common topographic characteristic in Korea is a mountainous region along with alluvial plain, and consequently it is difficult to apply previous algorithms directly to this area. A new algorithm of automatic lineament extraction from remotely sensed images is developed in this study specifically for geological applications. An algorithm, named as DSTA(Dynamic Segment Tracing Algorithm), is developed to produce binary image composed of linear component and non-linear component. The proposed algorithm effectively reduces the look direction bias associated with sun's azimuth angle and the noise in the low contrast region by utilizing a dynamic sub window. This algorithm can successfully accomodate lineaments in the alluvial plain as well as mountainous region. Two additional algorithms for estimating the individual lineament vector, named as ALEHHT(Automatic Lineament Extraction by Hierarchical Hough Transform) and ALEGHT(Automatic Lineament Extraction by Generalized Hough Transform) which are merging operation steps through the Hierarchical Hough transform and Generalized Hough transform respectively, are also developed to generate geological lineaments. The merging operation proposed in this study is consisted of three parameters: the angle between two lines($\delta$$\beta$), the perpendicular distance($(d_ij)$), and the distance between midpoints of lines(dn). The test result of the developed algorithm using Landsat TM image demonstrates that lineaments in alluvial plain as well as in rugged mountain is extremely well extracted. Even the lineaments parallel to sun's azimuth angle are also well detected by this approach. Further study is, however, required to accommodate the effect of quantization interval(droh) parameter in ALEGHT for optimization.

A Time Series Graph based Convolutional Neural Network Model for Effective Input Variable Pattern Learning : Application to the Prediction of Stock Market (효과적인 입력변수 패턴 학습을 위한 시계열 그래프 기반 합성곱 신경망 모형: 주식시장 예측에의 응용)

  • Lee, Mo-Se;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.167-181
    • /
    • 2018
  • Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN(Convolutional Neural Network), which is known as the effective solution for recognizing and classifying images or voices, has been popularly applied to classification and prediction problems. In this study, we investigate the way to apply CNN in business problem solving. Specifically, this study propose to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. As mentioned, CNN has strength in interpreting images. Thus, the model proposed in this study adopts CNN as the binary classifier that predicts stock market direction (upward or downward) by using time series graphs as its inputs. That is, our proposal is to build a machine learning algorithm that mimics an experts called 'technical analysts' who examine the graph of past price movement, and predict future financial price movements. Our proposed model named 'CNN-FG(Convolutional Neural Network using Fluctuation Graph)' consists of five steps. In the first step, it divides the dataset into the intervals of 5 days. And then, it creates time series graphs for the divided dataset in step 2. The size of the image in which the graph is drawn is $40(pixels){\times}40(pixels)$, and the graph of each independent variable was drawn using different colors. In step 3, the model converts the images into the matrices. Each image is converted into the combination of three matrices in order to express the value of the color using R(red), G(green), and B(blue) scale. In the next step, it splits the dataset of the graph images into training and validation datasets. We used 80% of the total dataset as the training dataset, and the remaining 20% as the validation dataset. And then, CNN classifiers are trained using the images of training dataset in the final step. Regarding the parameters of CNN-FG, we adopted two convolution filters ($5{\times}5{\times}6$ and $5{\times}5{\times}9$) in the convolution layer. In the pooling layer, $2{\times}2$ max pooling filter was used. The numbers of the nodes in two hidden layers were set to, respectively, 900 and 32, and the number of the nodes in the output layer was set to 2(one is for the prediction of upward trend, and the other one is for downward trend). Activation functions for the convolution layer and the hidden layer were set to ReLU(Rectified Linear Unit), and one for the output layer set to Softmax function. To validate our model - CNN-FG, we applied it to the prediction of KOSPI200 for 2,026 days in eight years (from 2009 to 2016). To match the proportions of the two groups in the independent variable (i.e. tomorrow's stock market movement), we selected 1,950 samples by applying random sampling. Finally, we built the training dataset using 80% of the total dataset (1,560 samples), and the validation dataset using 20% (390 samples). The dependent variables of the experimental dataset included twelve technical indicators popularly been used in the previous studies. They include Stochastic %K, Stochastic %D, Momentum, ROC(rate of change), LW %R(Larry William's %R), A/D oscillator(accumulation/distribution oscillator), OSCP(price oscillator), CCI(commodity channel index), and so on. To confirm the superiority of CNN-FG, we compared its prediction accuracy with the ones of other classification models. Experimental results showed that CNN-FG outperforms LOGIT(logistic regression), ANN(artificial neural network), and SVM(support vector machine) with the statistical significance. These empirical results imply that converting time series business data into graphs and building CNN-based classification models using these graphs can be effective from the perspective of prediction accuracy. Thus, this paper sheds a light on how to apply deep learning techniques to the domain of business problem solving.