• Title/Summary/Keyword: 3D data extraction

Search Result 330, Processing Time 0.028 seconds

Twitter Issue Tracking System by Topic Modeling Techniques (토픽 모델링을 이용한 트위터 이슈 트래킹 시스템)

  • Bae, Jung-Hwan;Han, Nam-Gi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.109-122
    • /
    • 2014
  • People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.

A Study on the Estimation of Multi-Object Social Distancing Using Stereo Vision and AlphaPose (Stereo Vision과 AlphaPose를 이용한 다중 객체 거리 추정 방법에 관한 연구)

  • Lee, Ju-Min;Bae, Hyeon-Jae;Jang, Gyu-Jin;Kim, Jin-Pyeong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.7
    • /
    • pp.279-286
    • /
    • 2021
  • Recently, We are carrying out a policy of physical distancing of at least 1m from each other to prevent the spreading of COVID-19 disease in public places. In this paper, we propose a method for measuring distances between people in real time and an automation system that recognizes objects that are within 1 meter of each other from stereo images acquired by drones or CCTVs according to the estimated distance. A problem with existing methods used to estimate distances between multiple objects is that they do not obtain three-dimensional information of objects using only one CCTV. his is because three-dimensional information is necessary to measure distances between people when they are right next to each other or overlap in two dimensional image. Furthermore, they use only the Bounding Box information to obtain the exact coordinates of human existence. Therefore, in this paper, to obtain the exact two-dimensional coordinate value in which a person exists, we extract a person's key point to detect the location, convert it to a three-dimensional coordinate value using Stereo Vision and Camera Calibration, and estimate the Euclidean distance between people. As a result of performing an experiment for estimating the accuracy of 3D coordinates and the distance between objects (persons), the average error within 0.098m was shown in the estimation of the distance between multiple people within 1m.

Investigation of Intertidal Zone using TerraSAR-X (TerraSAR-X를 이용한 조간대 관측)

  • Park, Jeong-Won;Lee, Yoon-Kyung;Won, Joong-Sun
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.4
    • /
    • pp.383-389
    • /
    • 2009
  • The main objective of the research is a feasibility study on the intertidal zone using a X-band radar satellite, TerraSAR-X. The TerraSAR-X data have been acquired in the west coast of Korea where large tidal flats, Ganghwa and Yeongjong tidal flats, are developed. Investigations include: 1) waterline and backscattering characteristics of the high resolution X-band images in tidal flats; 2) polarimetric signature of halophytes (or salt marsh plants), specifically Suaeda japonica; and 3) phase and coherence of interferometric pairs. Waterlines from TerraSAR-X data satisfy the requirement of horizontal accuracy of 60 m that corresponds to 20 cm in average height difference while current other spaceborne SAR systems could not meet the requirement. HH-polarization was the best for extraction of waterline, and its geometric position is reliable due to the short wavelength and accurate orbit control of the TerraSAR-X. A halophyte or salt marsh plant, Suaeda japonica, is an indicator of local sea level change. From X-band ground radar measurements, a dual polarization of VV/VH-pol. is anticipated to be the best for detection of the plant with about 9 dB difference at 35 degree incidence angle. However, TerraSAR-X HH/TV dual polarization was turned to be more effective for salt marsh monitoring. The HH-HV value was the maximum of about 7.9 dB at 31.6 degree incidence angle, which is fairly consistent with the results of X-band ground radar measurement. The boundary of salt marsh is effectively traceable specifically by TerraSAR-X cross-polarization data. While interferometric phase is not coherent within normal tidal flat, areas of salt marsh where the landization is preceded show coherent interferometric phases regardless of seasons or tide conditions. Although TerraSAR-X interferometry may not be effective to directly measure height or changes in tidal flat surface, TanDEM-X or other future X-band SAR tandem missions within one-day interval would be useful for mapping tidal flat topography.

Production and Action of Microbial Piscicidal Substance (미생물에 의한 살어성물질의 생성 및 그 작용)

  • 도재호;서정훈
    • Microbiology and Biotechnology Letters
    • /
    • v.6 no.1
    • /
    • pp.41-46
    • /
    • 1978
  • Piscicidal substance produced by Streptomyces sp. isolated from soil was toxic against various kinds of fish. After extraction with CH$Cl_3$ from the culture medium, the substance was purified by avicel column chromatography. In order to test toxicity, various kinds of fish were subjected to the acqueous solution of 100 us of the substance per liter of water. Generally, the substance was toxic to most fish, but Macropodus chinenes and Misgurnus mizolepis are resistant to the substance than Gobius similis and Pseudorasbora parva. The substance was stable at pH range, 3.0 to 7.0, but labile at alkaline pH, and to heat as well. Succinic dehydrogenase on most of tissue cell of Cyprinus carpio was inhibited by this substance strongly, but spinal cord was not inhibited. By addition of Cu and Pb salts to the culture medium, piscicidal substance producibility was activated.

  • PDF

Development of AAB (Algorithm-Aided BIM) Based 3D Design Bases Management System in Nuclear Power Plant (Algorithm-Aided BIM 기반 원전 3차원 설계기준 관리시스템 개발)

  • Shin, Jaeseop
    • Korean Journal of Construction Engineering and Management
    • /
    • v.20 no.2
    • /
    • pp.28-36
    • /
    • 2019
  • The APR1400 (Advanced Power Reactor 1400MW) nuclear power plant is a large-scale national infrastructure facility with a total project cost of 8.6 trillion won and a project period of 10 years or more. The total project area is about 2.17 million square meters and consists of more than 20 buildings and structures. And the total number of drawings required for construction is about 65,000. In order to design such a large facility, it is important to establish a design standard that reflects the design intent and can increase conformity between documents (drawings). To this end, a design bases document (DBD) reflecting the design bases that extracted in regulatory requirements (e.g. 10CFR50, Korean Law, etc.) is created. However, although the design bases are important concepts that are a big framework for the whole design of the nuclear power plant, they are managed in 2-dimensional by the experts in each field fragmentarily. Therefore, in order to improve the usability of building information, we developed BIM(Building Information Model) based 3-dimensional design bases management system. For this purpose, the concept of design bases information layer (DBIL) was introduced. Through the simulation of developed system, design bases attribute and element data extraction for each DBIL was confirmed, and walls, floors, doors, and penetrations with DBIL were successfully extracted.

Analysis of Mineral and Volatile Flavor Compounds in Pimpinella brachycarpa N. by ICP-AES and SDE, HS-SPME-GC/MS (ICP-AES와 SDE, HS-SPME-GC/MS를 이용한 참나물의 무기성분과 향기성분)

  • Chang, Kyung-Mi;Chung, Mi-Sook;Kim, Mi-Kyung;Kim, Gun-Hee
    • Journal of the Korean Society of Food Culture
    • /
    • v.22 no.2
    • /
    • pp.246-253
    • /
    • 2007
  • Mineral and volatile flavor compounds of Pimpinella brochycarpa N., a perennial Korean medicinal plant of the Umbelliferae family, were analyzed by inductively coupled plasma-atomic emission spectroscopy (ICP-AES) and simultaneous steam distillation extract (SDE)-gas chromatography mass spectrometry (GC/MS), head space solid phase micro-extraction (HS-SPME)-GC/MS. Mineral contents of the stalks and leaves were compared and the flavor patterns of the fresh and the shady air-dried samples were obtained by the electronic nose (EN) with 6 metal oxide sensors. Principal component analysis (PCA) was carried out using the data obtained from EN. The 1st principal values of the fresh samples have + values and the shady air-dried have - values. The essential oil extracted from the fresh and the shady air-dried by SDE method contain 58 and 31 flavor compounds. When HS-SPME method with CAR/PDMS fiber and PDMS fiber were used, 34 and 21 flavor compounds. The principal volatile components of Pimpinella brachycarpa N. were ${\alpha}$-selinene, germacrene D, and myrcene.

Gene Expression Biodosimetry: Quantitative Assessment of Radiation Dose with Total Body Exposure of Rats

  • Saberi, Alihossein;Khodamoradi, Ehsan;Birgani, Mohammad Javad Tahmasebi;Makvandi, Manoochehr
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.16 no.18
    • /
    • pp.8553-8557
    • /
    • 2016
  • Background: Accurate dose assessment and correct identification of irradiated from non-irradiated people are goals of biological dosimetry in radiation accidents. Objectives: Changes in the FDXR and the RAD51 gene expression (GE) levels were here analyzed in response to total body exposure (TBE) to a 6 MV x-ray beam in rats. We determined the accuracy for absolute quantification of GE to predict the dose at 24 hours. Materials and Methods: For this in vivo experimental study, using simple randomized sampling, peripheral blood samples were collected from a total of 20 Wistar rats at 24 hours following exposure of total body to 6 MV X-ray beam energy with doses (0.2, 0.5, 2 and 4 Gy) for TBE in Linac Varian 2100C/D (Varian, USA) in Golestan Hospital, in Ahvaz, Iran. Also, 9 rats was irradiated with a 6MV X-ray beam at doses of 1, 2, 3 Gy in 6MV energy as a validation group. A sham group was also included. After RNA extraction and DNA synthesis, GE changes were measured by the QRT-PCR technique and an absolute quantification strategy by taqman methodology in peripheral blood from rats. ROC analysis was used to distinguish irradiated from non-irradiated samples (qualitative dose assessment) at a dose of 2 Gy. Results: The best fits for mean of responses were polynomial equations with a R2 of 0.98 and 0.90 (for FDXR and RAD51 dose response curves, respectively). Dose response of the FDXR gene produced a better mean dose estimation of irradiated "validation" samples compared to the RAD51 gene at doses of 1, 2 and 3 Gy. FDXR gene expression separated the irradiated rats from controls with a sensitivity, specificity and accuracy of 87.5%, 83.5% and 81.3%, respectively, 24 hours after dose of 2 Gy. These values were significantly (p<0.05) higher than the 75%, 75% and 75%, respectively, obtained using gene expression of RAD51 analysis at a dose of 2 Gy. Conclusions: Collectively, these data suggest that absolute quantification by gel purified quantitative RT-PCR can be used to measure the mRNA copies for GE biodosimetry studies at comparable accuracy to similar methods. In the case of TBE with 6MV energy, FDXR gene expression analysis is more precise than that with RAD51 for quantitative and qualitative dose assessment.

Development of Rapid and Simple Drug Identification and Semi Quantitative Analytical Program by Gas Chromatography-Mass Spectrometry (가스크로마토그래피/질량분석기를 이용한 약물의 확인 및 간이 정량분석 프로그램 개발)

  • Kim, Eun-Mi;Han, Eun-Young;Hong, Hyo-Jeong;Jeong, Su-Jin;Choe, Sang-Gil;Rhee, Jong-Sook;Jung, Jin-Mi;Yeom, Hye-Sun;Lee, Han-Sun;Lee, Sang-Ki
    • YAKHAK HOEJI
    • /
    • v.55 no.2
    • /
    • pp.106-115
    • /
    • 2011
  • Systematic toxicological analysis (STA) means the process for general unknown screening of drugs and toxic compounds in biological fluids. In order to establish STA, in previous study we investigated pattern of drugs & poisons in autopsy cases during 2007~2009 in Korea, and finally selected 62 drugs as target drugs for STA. In this study, rapid and simple drug identification and quantitative analytical program by gas chromatography-mass spectrometry(GC-MS) was developed. The in-house program, "DrugMan", consisted of modified chemstation data analysis menu and newly developed macro modules. Total 55 drugs among 62 target drugs were applied to this program, they were 14 antidepressants, 8 anti-histamines, 5 sedatives/hypnotics, 5 narcotic analgesics, 3 antipsychotic drugs, and etc. For calibration curves, fifty five drugs were divided into four groups of range considering their therapeutic or toxic concentrations in blood specimen, i.e. 0.05~1 mg/l, 0.1~1 mg/l, 0.1~5 mg/l or 0.5~10 mg/l. Standards spiked bloods were extracted by solid-phase extraction (SPE) with trimipramine-D3 as internal standard. Parameters such as retention times, 3 mass fragment ions, and calibration curves for each drug were registered to DrugMan. A series of identification, semi quantitation of target drugs and reporting the results were performed automatically. Calibration curves for most drugs were linear with correlation coefficients exceeding 0.98. Sensitivity rate of DrugMan was 0.90 (90%) for 55 drugs at the level of 0.5 mg/l. For standard spiked bloods at the level of 0.5 mg/l for 29 drugs, semi quantitative concentrations were ranged 0.36~0.64 mg/l by DrugMan. If more drugs are registered to database in DrugMan in further study, it will be useful tools for STA in forensic toxicology.

Application and Validation of an Optimal Analytical Method using QuEChERS for the determination of Tolpyralate in Agricultural Products (QuEChERS법을 활용한 농산물 중 제초제 Tolpyralate의 최적 분석법 선발 및 검증)

  • Lee, Han Sol;Park, Ji-Su;Lee, Su Jung;Shin, Hye-Sun;Kim, Ji-Young;Yun, Sang Soon;Jung, Yong-hyun;Oh, Jae-Ho
    • Korean Journal of Environmental Agriculture
    • /
    • v.39 no.3
    • /
    • pp.246-252
    • /
    • 2020
  • BACKGROUND: Pesticides are broadly used to control weeds and pests, and the residues remaining in crops are managed in accordance with the MRLs (maximum residue limits). Therefore, an analytical method is required to quantify the residues, and we conducted a series of analyses to select and validate the quick and simple analytical method for tolpyralate in five agricultural products using QuEChERS (quick, easy, cheap, effective, rugged and safe) method and LC-MS/MS (liquid chromatography-tandem mass spectrometry). METHODS AND RESULTS: The agricultural samples were extracted with acetonitrile followed by addition of anhydrous magnesium sulfate, sodium chloride, disodium hydrogencitrate sesquihydrate and trisodium citrate dihydrate. After shaking and centrifugation, purification was performed with d-SPE (dispersive-solid phase extraction) sorbents. To validate the optimized method, its selectivity, linearity, LOD (limit of detection), LOQ (limit of quantitation), accuracy, repeatability, and reproducibility from the inter-laboratory analyses were considered. LOQ of the analytical method was 0.01 mg/kg at five agricultural products and the linearity of matrix-matched calibration were good at seven concentration levels, from 0.0025 to 0.25 mg/L (R2≥0.9980). Mean recoveries at three spiking levels (n=5) were in the range of 85.2~112.4% with associated relative standard deviation values less than 6.2%, and the coefficient of variation between the two laboratories was also below 13%. All optimized results were validated according to the criteria ranges requested in the Codex Alimentarius Commission (CAC) and Ministry of Food and Drug Safety (MFDS) guidelines. CONCLUSION: In conclusion, we suggest that the selected and validated method could serve as a basic data for detecting tolpyralate residue in imported and domestic agricultural products.

Label Embedding for Improving Classification Accuracy UsingAutoEncoderwithSkip-Connections (다중 레이블 분류의 정확도 향상을 위한 스킵 연결 오토인코더 기반 레이블 임베딩 방법론)

  • Kim, Museong;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.175-197
    • /
    • 2021
  • Recently, with the development of deep learning technology, research on unstructured data analysis is being actively conducted, and it is showing remarkable results in various fields such as classification, summary, and generation. Among various text analysis fields, text classification is the most widely used technology in academia and industry. Text classification includes binary class classification with one label among two classes, multi-class classification with one label among several classes, and multi-label classification with multiple labels among several classes. In particular, multi-label classification requires a different training method from binary class classification and multi-class classification because of the characteristic of having multiple labels. In addition, since the number of labels to be predicted increases as the number of labels and classes increases, there is a limitation in that performance improvement is difficult due to an increase in prediction difficulty. To overcome these limitations, (i) compressing the initially given high-dimensional label space into a low-dimensional latent label space, (ii) after performing training to predict the compressed label, (iii) restoring the predicted label to the high-dimensional original label space, research on label embedding is being actively conducted. Typical label embedding techniques include Principal Label Space Transformation (PLST), Multi-Label Classification via Boolean Matrix Decomposition (MLC-BMaD), and Bayesian Multi-Label Compressed Sensing (BML-CS). However, since these techniques consider only the linear relationship between labels or compress the labels by random transformation, it is difficult to understand the non-linear relationship between labels, so there is a limitation in that it is not possible to create a latent label space sufficiently containing the information of the original label. Recently, there have been increasing attempts to improve performance by applying deep learning technology to label embedding. Label embedding using an autoencoder, a deep learning model that is effective for data compression and restoration, is representative. However, the traditional autoencoder-based label embedding has a limitation in that a large amount of information loss occurs when compressing a high-dimensional label space having a myriad of classes into a low-dimensional latent label space. This can be found in the gradient loss problem that occurs in the backpropagation process of learning. To solve this problem, skip connection was devised, and by adding the input of the layer to the output to prevent gradient loss during backpropagation, efficient learning is possible even when the layer is deep. Skip connection is mainly used for image feature extraction in convolutional neural networks, but studies using skip connection in autoencoder or label embedding process are still lacking. Therefore, in this study, we propose an autoencoder-based label embedding methodology in which skip connections are added to each of the encoder and decoder to form a low-dimensional latent label space that reflects the information of the high-dimensional label space well. In addition, the proposed methodology was applied to actual paper keywords to derive the high-dimensional keyword label space and the low-dimensional latent label space. Using this, we conducted an experiment to predict the compressed keyword vector existing in the latent label space from the paper abstract and to evaluate the multi-label classification by restoring the predicted keyword vector back to the original label space. As a result, the accuracy, precision, recall, and F1 score used as performance indicators showed far superior performance in multi-label classification based on the proposed methodology compared to traditional multi-label classification methods. This can be seen that the low-dimensional latent label space derived through the proposed methodology well reflected the information of the high-dimensional label space, which ultimately led to the improvement of the performance of the multi-label classification itself. In addition, the utility of the proposed methodology was identified by comparing the performance of the proposed methodology according to the domain characteristics and the number of dimensions of the latent label space.