• Title/Summary/Keyword: Big Data Structure

Search Result 386, Processing Time 0.032 seconds

Inter-category Map: Building Cognition Network of General Customers through Big Data Mining

  • Song, Gil-Young;Cheon, Youngjoon;Lee, Kihwang;Park, Kyung Min;Rim, Hae-Chang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.2
    • /
    • pp.583-600
    • /
    • 2014
  • Social media is considered a valuable platform for gathering and analyzing the collective and subconscious opinions of people in Internet and mobile environments, where they express, explicitly and implicitly, their daily preferences for brands and products. Extracting and tracking the various attitudes and concerns that people express through social media could enable us to categorize brands and decipher individuals' cognitive decision-making structure in their choice of brands. We investigate the cognitive network structure of consumers by building an inter-category map through the mining of big data. In so doing, we create an improved online recommendation model. Building on economic sociology theory, we suggest a framework for revealing collective preference by analyzing the patterns of brand names that users frequently mention in the online public sphere. We expect that our study will be useful for those conducting theoretical research on digital marketing strategies and doing practical work on branding strategies.

Detection of Malicious PDF based on Document Structure Features and Stream Objects

  • Kang, Ah Reum;Jeong, Young-Seob;Kim, Se Lyeong;Kim, Jonghyun;Woo, Jiyoung;Choi, Sunoh
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.11
    • /
    • pp.85-93
    • /
    • 2018
  • In recent years, there has been an increasing number of ways to distribute document-based malicious code using vulnerabilities in document files. Because document type malware is not an executable file itself, it is easy to bypass existing security programs, so research on a model to detect it is necessary. In this study, we extract main features from the document structure and the JavaScript contained in the stream object In addition, when JavaScript is inserted, keywords with high occurrence frequency in malicious code such as function name, reserved word and the readable string in the script are extracted. Then, we generate a machine learning model that can distinguish between normal and malicious. In order to make it difficult to bypass, we try to achieve good performance in a black box type algorithm. For an experiment, a large amount of documents compared to previous studies is analyzed. Experimental results show 98.9% detection rate from three different type algorithms. SVM, which is a black box type algorithm and makes obfuscation difficult, shows much higher performance than in previous studies.

A Study on Veracity of Raw Data based on Value Creation -Focused on YouTube Monetization

  • CHOI, Seoyeon;SHIN, Seung-Jung
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.2
    • /
    • pp.218-223
    • /
    • 2021
  • The five elements of big data are said to be Volume, Variety, Velocity, Veracity, and Value. Among them, data lacking the Veracity of the data or fake data not only makes an error in decision making, but also hinders the creation of value. This study analyzed YouTube's revenue structure to focus the effect of data integrity on data valuation among these five factors. YouTube is one of the OTT service platforms, and due to COVID-19 in 2020, YouTube creators have emerged as a new profession. Among the revenue-generating models provided by YouTube, the process of generating advertising revenue based on click-based playback was analyzed. And, analyzed the process of subtracting the profits generated from invalid activities that not the clicks due to viewers' pure interests, then paying the final revenue. The invalid activity in YouTube's revenue structure is Raw Data, not pure viewing activity of viewers, and it was confirmed a direct impact on revenue generation. Through the analysis of this process, the new Data Value Chain was proposed.

Strategies of Knowledge Pricing and the Impact on Firms' New Product Development Performance

  • Wu, Chuanrong;Tan, Ning;Lu, Zhi;Yang, Xiaoming;McMurtrey, Mark E.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.3068-3085
    • /
    • 2021
  • The economics of big data knowledge, especially cloud computing and statistical data of consumer preferences, has attracted increasing academic and industry practitioners' attention. Firms nowadays require purchasing not only external private patent knowledge from other firms, but also proprietary big data knowledge to support their new product development. Extant research investigates pricing strategies of external private patent knowledge and proprietary big data knowledge separately. Yet, a comprehensive investigation of pricing strategies of these two types of knowledge is in pressing need. This research constructs an overarching pricing model of external private patent knowledge and proprietary big data knowledge through the lens of firm profitability as a knowledge transaction recipient. The proposed model can help those firms who purchase external knowledge choose the optimal knowledge structure and pricing strategies of two types of knowledge, and provide theoretical and methodological guidance for knowledge transaction recipient firms to negotiate with knowledge providers.

A Study on the Platform for Big Data Analysis of Manufacturing Process (제조 공정 빅데이터 분석을 위한 플랫폼 연구)

  • Ku, Jin-Hee
    • Journal of Convergence for Information Technology
    • /
    • v.7 no.5
    • /
    • pp.177-182
    • /
    • 2017
  • As major ICT technologies such as IoT, cloud computing, and Big Data are being applied to manufacturing, smart factories are beginning to be built. The key of smart factory implementation is the ability to acquire and analyze data of the factory. Therefore, the need for a big data analysis platform is increasing. The purpose of this study is to construct a platform for big data analysis of manufacturing process and propose integrated method for analysis. The proposed platform is a RHadoop-based structure that integrates analysis tool R and Hadoop to distribute a large amount of datasets. It can store and analyze big data collected in the unit process and factory in the automation system directly in HBase, and it has overcome the limitations of RDB - based analysis. Such a platform should be developed in consideration of the unit process suitability for smart factories, and it is expected to be a guide to building IoT platforms for SMEs that intend to introduce smart factories into the manufacturing process.

Designing Bigdata Platform for Multi-Source Maritime Information

  • Junsang Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.1
    • /
    • pp.111-119
    • /
    • 2024
  • In this paper, we propose a big data platform that can collect information from various sources collected at ocean. Currently operating ocean-related big data platforms are focused on storing and sharing created data, and each data provider is responsible for data collection and preprocessing. There are high costs and inefficiencies in collecting and integrating data in a marine environment using communication networks that are poor compared to those on land, making it difficult to implement related infrastructure. In particular, in fields that require real-time data collection and analysis, such as weather information, radar and sensor data, a number of issues must be considered compared to land-based systems, such as data security, characteristics of organizations and ships, and data collection costs, in addition to communication network issues. First, this paper defines these problems and presents solutions. In order to design a big data platform that reflects this, we first propose a data source, hierarchical MEC, and data flow structure, and then present an overall platform structure that integrates them all.

Dynamic Personal Knowledge Network Design based on Correlated Connection Structure (결합 연결구조 기반의 동적 개인 지식네트워크 설계)

  • Shim, JeongYon
    • The Journal of Korean Association of Computer Education
    • /
    • v.18 no.6
    • /
    • pp.71-79
    • /
    • 2015
  • In a new era of Cloud and Big data, how to search the useful data from dynamic huge data pool in a right time and right way is most important at the stage where the information is getting more important. Above all, in the era of s Big Data it is required to design the advanced efficient intelligent Knowledge system which can process the dynamic variable big data. Accordingly in this paper we propose Dynamic personal Knowledge Network as one of the advanced Intelligent system approach. Adopting the human brain function and its neuro dynamics, an Intelligent system which has a structural flexibility was designed. For Structure-Function association, a personal Knowledge Network is made to be structured and to have reorganizing function as connecting the common nodes. We also design this system to have a reasoning process in the extracted optimal paths from the Knowledge Network.

Attention-based word correlation analysis system for big data analysis (빅데이터 분석을 위한 어텐션 기반의 단어 연관관계 분석 시스템)

  • Chi-Gon, Hwang;Chang-Pyo, Yoon;Soo-Wook, Lee
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.27 no.1
    • /
    • pp.41-46
    • /
    • 2023
  • Recently, big data analysis can use various techniques according to the development of machine learning. Big data collected in reality lacks an automated refining technique for the same or similar terms based on semantic analysis of the relationship between words. Since most of the big data is described in general sentences, it is difficult to understand the meaning and terms of the sentences. To solve these problems, it is necessary to understand the morphological analysis and meaning of sentences. Accordingly, NLP, a technique for analyzing natural language, can understand the word's relationship and sentences. Among the NLP techniques, the transformer has been proposed as a way to solve the disadvantages of RNN by using self-attention composed of an encoder-decoder structure of seq2seq. In this paper, transformers are used as a way to form associations between words in order to understand the words and phrases of sentences extracted from big data.

A Study on the Anomaly Prediction System of Drone Using Big Data (빅데이터를 활용한 드론의 이상 예측시스템 연구)

  • Lee, Yang-Kyoo;Hong, Jun-Ki;Hong, Sung-Chan
    • Journal of Internet Computing and Services
    • /
    • v.21 no.2
    • /
    • pp.27-37
    • /
    • 2020
  • Recently, big data is rapidly emerging as a core technology in the 4th industrial revolution. Further, the utilization and the demand of drones are continuously increasing with the development of the 4th industrial revolution. However, as the drones usage increases, the risk of drones falling increases. Drones always have a risk of being able to fall easily even with small problems due to its simple structure. In this paper, in order to predict the risk of drone fall and to prevent the fall, ESC (Electronic Speed Control) is attached integrally with the drone's driving motor and the acceleration sensor is stored to collect the vibration data in real time. By processing and monitoring the data in real time and analyzing the data through big data obtained in such a situation using a Fast Fourier Transform (FFT) algorithm, we proposed a prediction system that minimizes the risk of drone fall by analyzing big data collected from drones.

A Study on Interdisciplinary Structure of Big Data Research with Journal-Level Bibliographic-Coupling Analysis (학술지 단위 서지결합분석을 통한 빅데이터 연구분야의 학제적 구조에 관한 연구)

  • Lee, Boram;Chung, EunKyung
    • Journal of the Korean Society for information Management
    • /
    • v.33 no.3
    • /
    • pp.133-154
    • /
    • 2016
  • Interdisciplinary approach has been recognized as one of key strategies to address various and complex research problems in modern science. The purpose of this study is to investigate the interdisciplinary characteristics and structure of the field of big data. Among the 1,083 journals related to the field of big data, multiple Subject Categories (SC) from the Web of Science were assigned to 420 journals (38.8%) and 239 journals (22.1%) were assigned with the SCs from different fields. These results show that the field of big data indicates the characteristics of interdisciplinarity. In addition, through bibliographic coupling network analysis of top 56 journals, 10 clusters in the network were recognized. Among the 10 clusters, 7 clusters were from computer science field focusing on technical aspects such as storing, processing and analyzing the data. The results of cluster analysis also identified multiple research works of analyzing and utilizing big data in various fields such as science & technology, engineering, communication, law, geography, bio-engineering and etc. Finally, with measuring three types of centrality (betweenness centrality, nearest centrality, triangle betweenness centrality) of journals, computer science journals appeared to have strong impact and subjective relations to other fields in the network.