• 제목/요약/키워드: Big data platform

검색결과 503건 처리시간 0.038초

지역중심의 스마트관광 생태계 지원 서비스 플랫 (Service Platform of Regional Smart Tour Ecosystem Support)

  • 원달수
    • 문화기술의 융합
    • /
    • 제4권4호
    • /
    • pp.31-36
    • /
    • 2018
  • 광 산업은 국가 경제 활성화에 지대한 영향력을 갖고 있으며, IT기술의 발전은 관광객의 특성, 행위, 구매 성향, 관심사 등에 기반한 개인 프로파일 정보 및 위치정보, 활동정보 등의 수집과 분석이 가능해졌다. 이를 구현하기 위해 융합형 스마트관광 정보 서비스 플랫폼 구현은 3단계로 나누어 비지니스 모델 개발, IoT & 빅데이터 통합관리 시스템, 빅데이터 알고리즘 개발 및 분석 플랫폼 개발로 완성된다. 플랫폼 및 알고리즘의 원천기술은 오픈소스를 채택하고 그 기반위에 서비스 요소를 확장한 후, 지역을 연계한 Test-Bed 실증 시험을 통해 문제점을 보완하는 과정을 진행하게 된다. 이 플랫폼을 활용하면 다양한 정보를 통합적으로 분석하여 관광객별 맞춤화된 서비스를 제공할 수 있는 스마트관광 환경이 가능해진다. 또한 지역중심의 스마트관광 생태계 조성을 통해 관광 목적지 주민의 삶을 개선하고 지역 재생과 일자리 창출에도 기여할 수가 있을 것이다.

하둡과 순차패턴 마이닝 기술을 통한 교통카드 빅데이터 분석 (Analysis of Traffic Card Big Data by Hadoop and Sequential Mining Technique)

  • 김우생;김용훈;박희성;박진규
    • Journal of Information Technology Applications and Management
    • /
    • 제24권4호
    • /
    • pp.187-196
    • /
    • 2017
  • It is urgent to prepare countermeasures for traffic congestion problems of Korea's metropolitan area where central functions such as economic, social, cultural, and education are excessively concentrated. Most users of public transportation in metropolitan areas including Seoul use the traffic cards. If various information is extracted from traffic big data produced by the traffic cards, they can provide basic data for transport policies, land usages, or facility plans. Therefore, in this study, we extract valuable information such as the subway passengers' frequent travel patterns from the big traffic data provided by the Seoul Metropolitan Government Big Data Campus. For this, we use a Hadoop (High-Availability Distributed Object-Oriented Platform) to preprocess the big data and store it into a Mongo database in order to analyze it by a sequential pattern data mining technique. Since we analysis the actual big data, that is, the traffic cards' data provided by the Seoul Metropolitan Government Big Data Campus, the analyzed results can be used as an important referenced data when the Seoul government makes a plan about the metropolitan traffic policies.

A Context-Awareness Modeling User Profile Construction Method for Personalized Information Retrieval System

  • Kim, Jee Hyun;Gao, Qian;Cho, Young Im
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • 제14권2호
    • /
    • pp.122-129
    • /
    • 2014
  • Effective information gathering and retrieval of the most relevant web documents on the topic of interest is difficult due to the large amount of information that exists in various formats. Current information gathering and retrieval techniques are unable to exploit semantic knowledge within documents in the "big data" environment; therefore, they cannot provide precise answers to specific questions. Existing commercial big data analytic platforms are restricted to a single data type; moreover, different big data analytic platforms are effective at processing different data types. Therefore, the development of a common big data platform that is suitable for efficiently processing various data types is needed. Furthermore, users often possess more than one intelligent device. It is therefore important to find an efficient preference profile construction approach to record the user context and personalized applications. In this way, user needs can be tailored according to the user's dynamic interests by tracking all devices owned by the user.

Study on Data Processing of the IOT Sensor Network Based on a Hadoop Cloud Platform and a TWLGA Scheduling Algorithm

  • Li, Guoyu;Yang, Kang
    • Journal of Information Processing Systems
    • /
    • 제17권6호
    • /
    • pp.1035-1043
    • /
    • 2021
  • An Internet of Things (IOT) sensor network is an effective solution for monitoring environmental conditions. However, IOT sensor networks generate massive data such that the abilities of massive data storage, processing, and query become technical challenges. To solve the problem, a Hadoop cloud platform is proposed. Using the time and workload genetic algorithm (TWLGA), the data processing platform enables the work of one node to be shared with other nodes, which not only raises efficiency of one single node but also provides the compatibility support to reduce the possible risk of software and hardware. In this experiment, a Hadoop cluster platform with TWLGA scheduling algorithm is developed, and the performance of the platform is tested. The results show that the Hadoop cloud platform is suitable for big data processing requirements of IOT sensor networks.

블록체인 네트워크를 이용한 소규모 분산전력 거래플랫폼의 정산소요시간에 관한 연구 (A Study on the Accounts Balancing Time of Small Distributed Power Trading Platform Using Block Chain Network)

  • 김영곤;허걸;최중인;위재우
    • 에너지공학
    • /
    • 제27권4호
    • /
    • pp.86-91
    • /
    • 2018
  • 이 논문은 블록체인[1] 기술을 활용한 소규모 분산전력자원 거래 플랫폼에서의 정산소요시간에 대한 고찰이다. 먼저 연구에 적용한 "AMI 인프라를 활용한 국민 VPP 에너지 관리 시스템 (AI 기반의 에너지 거래 플랫폼)"을 소개한 후, 테스트베드 환경 내 IoT 전력 빅데이터[2] 분석으로 인증된 프로슈머의 발전(감축)량에 근거하여 지급되는 블록체인 암호화폐 코인의 정산과정 그리고 소요시간에 대하여 알아본다. 더불어 기존 람다 아키텍처에 MapD[3]를 적용한 GPU Fast 빅데이터 전력 빅데이터 분석 시스템 구성을 제시 한다.

빅데이터를 통한 OTT 오리지널 콘텐츠의 성공요인 분석, 넷플릭스의 '오징어게임 시즌2' 제언 (Analysis of Success Factors of OTT Original Contents Through BigData, Netflix's 'Squid Game Season 2' Proposal)

  • 안성훈;정재우;오세종
    • 디지털산업정보학회논문지
    • /
    • 제18권1호
    • /
    • pp.55-64
    • /
    • 2022
  • This study analyzes the success factors of OTT original content through big data, and intends to suggest scenarios, casting, fun, and moving elements when producing the next work. In addition, I would like to offer suggestions for the success of 'Squid Game Season 2'. The success factor of 'Squid Game' through big data is first, it is a simple psychological experimental game. Second, it is a retro strategy. Third, modern visual beauty and color. Fourth, it is simple aesthetics. Fifth, it is the platform of OTT Netflix. Sixth, Netflix's video recommendation algorithm. Seventh, it induced Binge-Watch. Lastly, it can be said that the consensus was high as it was related to the time to think about 'death' and 'money' in a pandemic situation. The suggestions for 'Squid Game Season 2' are as follows. First, it is a fusion of famous traditional games of each country. Second, it is an AI-based planned MD product production and sales strategy. Third, it is casting based on artificial intelligence big data. Fourth, secondary copyright and copyright sales strategy. The limitations of this study were analyzed only through external data. Data inside the Netflix platform was not utilized. In this study, if AI big data is used not only in the OTT field but also in entertainment and film companies, it will be possible to discover better business models and generate stable profits.

빅데이터 분석을 이용한 해양 구조물 배관 자재의 소요량 예측 (Estimation of Material Requirement of Piping Materials in an Offshore Structure using Big Data Analysis)

  • 오민재;노명일;박성우;김성훈
    • 대한조선학회논문집
    • /
    • 제55권3호
    • /
    • pp.243-251
    • /
    • 2018
  • In the shipyard, a lot of data is generated, stored, and managed during design, construction, and operation phases to build ships and offshore structures. However, it is difficult to handle such big data efficiently using existing data-handling technologies. As the big data technology is developed, the ship and offshore industries start to focus on the existing big data to find valuable information from it. In this paper, the material requirement estimation method of offshore structure piping materials using big data analysis is proposed. A big data platform for the data analysis in the shipyard is introduced and it is applied to the analysis of material requirement estimation to solve the problems in piping design by a designer. The regression model is developed from the big data of piping materials and verified using the existing data. This analysis can help a piping designer to estimate the exact amount of material requirement and schedule the purchase time.

A Study on Deep Learning Model-based Object Classification for Big Data Environment

  • Kim, Jeong-Sig;Kim, Jinhong
    • 한국소프트웨어감정평가학회 논문지
    • /
    • 제17권1호
    • /
    • pp.59-66
    • /
    • 2021
  • Recently, conceptual information model is changing fast, and these changes are coming about as a result of individual tendency, social cultural, new circumstances and societal shifts within big data environment. Despite the data is growing more and more, now is the time to commit ourselves to the development of renewable, invaluable information of social/live commerce. Because we have problems with various insoluble data, we propose about deep learning prediction model-based object classification in social commerce of big data environment. Accordingly, it is an increased need of social commerce platform capable of handling high volumes of multiple items by users. Consequently, responding to rapid changes in users is a very significant by deep learning. Namely, promptly meet the needs of the times, and a widespread growth in big data environment with the goal of realizing in this paper.

SNS 빅데이터 분석을 통한 재생에너지 동향 및 관계구조 (Renewable energy trends and relationship structure by SNS big data analysis)

  • 김종민
    • 융합보안논문지
    • /
    • 제22권1호
    • /
    • pp.55-60
    • /
    • 2022
  • 본 연구는 재생에너지와 관련된 에너지 분야의 동향과 관계구조를 분석하는 것이다. 이를 위해 본 연구에서는 SNS Data를 포한한 Big Data를 중점으로 분석하였다. SNS는 Instragram 플랫폼을 활용하여 재생에너지 해시태그들을 수집하였으며, 빅데이터 분석, 소셜네트워크 분석을 위한 워드임베딩 방법으로 사용하였고, 본 연구에서 도출된 결과를 토대로 재생에너지 산업의 발전에 활용할 수 있을 것으로 기대된다.

도커 기반의 실시간 데이터 연계 및 처리 환경을 고려한 빅데이터 관리 플랫폼 개발 (Development of Big-data Management Platform Considering Docker Based Real Time Data Connecting and Processing Environments)

  • 김동길;박용순;정태윤
    • 대한임베디드공학회논문지
    • /
    • 제16권4호
    • /
    • pp.153-161
    • /
    • 2021
  • Real-time access is required to handle continuous and unstructured data and should be flexible in management under dynamic state. Platform can be built to allow data collection, storage, and processing from local-server or multi-server. Although the former centralize method is easy to control, it creates an overload problem because it proceeds all the processing in one unit, and the latter distributed method performs parallel processing, so it is fast to respond and can easily scale system capacity, but the design is complex. This paper provides data collection and processing on one platform to derive significant insights from various data held by an enterprise or agency in the latter manner, which is intuitively available on dashboards and utilizes Spark to improve distributed processing performance. All service utilize dockers to distribute and management. The data used in this study was 100% collected from Kafka, showing that when the file size is 4.4 gigabytes, the data processing speed in spark cluster mode is 2 minute 15 seconds, about 3 minutes 19 seconds faster than the local mode.