Designing an Intelligent Advertising Business Model in Seoul's Metro Network (서울지하철의 지능형 광고 비즈니스모델 설계)
-
- Journal of Intelligence and Information Systems
- /
- v.23 no.4
- /
- pp.1-31
- /
- 2017
Modern businesses are adopting new technologies to serve their markets better as well as to improve efficiency and productivity. The advertising industry has continuously experienced disruptions from the traditional channels (radio, television and print media) to new complex ones including internet, social media and mobile-based advertising. This case study focuses on proposing intelligent advertising business model in Seoul's metro network. Seoul has one of the world's busiest metro network and transports a huge number of travelers on a daily basis. The high number of travelers coupled with a well-planned metro network creates a platform where marketers can initiate engagement and interact with both customers and potential customers. In the current advertising model, advertising is on illuminated and framed posters in the stations and in-car, non-illuminated posters, and digital screens that show scheduled arrivals and departures of metros. Some stations have digital screens that show adverts but they do not have location capability. Most of the current advertising media have one key limitation: space. For posters whether illuminated or not, one space can host only one advert at a time. Empirical literatures show that there is room for improving this advertising model and eliminate the space limitation by replacing the poster adverts with digital advertising platform. This new model will not only be digital, but will also provide intelligent advertising platform that is driven by data. The digital platform will incorporate location sensing, e-commerce, and mobile platform to create new value to all stakeholders. Travel cards used in the metro will be registered and the card scanners will have a capability to capture traveler's data when travelers tap their cards. This data once analyzed will make it possible to identify different customer groups. Advertisers and marketers will then be able to target specific customer groups, customize adverts based on the targeted consumer group, and offer a wide variety of advertising formats. Format includes video, cinemagraphs, moving pictures, and animation. Different advert formats create different emotions in the customer's mind and the goal should be to use format or combination of formats that arouse the expected emotion and lead to an engagement. Combination of different formats will be more effective and this can only work in a digital platform. Adverts will be location based, ensuring that adverts will show more frequently when the metro is near the premises of an advertiser. The advertising platform will automatically detect the next station and screens inside the metro will prioritize adverts in the station where the metro will be stopping. In the mobile platform, customers who opt to receive notifications will receive them when they approach the business premises of advertiser. The mobile platform will have indoor navigation for the underground shopping malls that will allow customers to search for facilities within the mall, products they may want to buy as well as deals going on in the underground mall. To create an end-to-end solution, the mobile solution will have a capability to allow customers purchase products through their phones, get coupons for deals, and review products and shops where they have bought a product. The indoor navigation will host intelligent mobile-based advertisement and a recommendation system. The indoor navigation will have adverts such that when a customer is searching for information, the recommendation system shows adverts that are near the place traveler is searching or in the direction that the traveler is moving. These adverts will be linked to the e-commerce platform such that if a customer clicks on an advert, it leads them to the product description page. The whole system will have multi-language as well as text-to-speech capability such that both locals and tourists have no language barrier. The implications of implementing this model are varied including support for small and medium businesses operating in the underground malls, improved customer experience, new job opportunities, additional revenue to business model operator, and flexibility in advertising. The new value created will benefit all the stakeholders.
Purpose: This paper proposes a performance evaluation method for state-of-the-art facility inspection/diagnostic equipment through a trend survey of equipment and standardization systems of US, Japan, and Korea. This paper also suggests the priority of developing a performance evaluation method through expert interviews and surveys. Method: In this study, report for the last 5 years of FMS, state-of-the-art equipment of facility maintenance companies/safety diagnosis specialist agencies and papers/research reports/patents of NTIS were analyzed to identify recent trends of facility inspection/diagnostic equipment usages. standardization system of US, Japan, and Korea were analyzed to figure out a suitable form of a performance evaluation method for the domestic situation. And expert interview and survey were conducted to identify the priority of developing a performance evaluation method. Result: The performance evaluation method must be developed by the shape that only evaluates performance, regardless of types of equipment, on inspection item level for creative technology development. The priority of developing the performance evaluation method was identified as crack detection of concrete for durability evaluation and displacement/deformation/fatigue detection of concrete and steel for stability evaluation. Conclusion: The performance evaluation method will be developed firstly for the crack detection of concrete for durability evaluation and displacement/deformation/fatigue detection of concrete/steel for stability evaluation. In order to promote creative technology development, the performance evaluation method should be developed in a form that provides standardized specimens or testbeds and can be applied regardless of types of technologies.
According as utilization for web increases rapidly, it is demanded that model about support interaction between web-based applications systematically and solutions can integrate new distributed platforms and existing environment effectively, accordingly, Web Services appeared by solution in reply. These days, a lot of software and hardware companies try to adoption of Web Services to their market, attenpt to construct their applications associationing components from various Web Services providers. However, to execute Web Services completely. it must have interoperability and need the standardization work that avoid thing which is subject to platform, application as well as service and programming language from other companies. WS-I (Web Services Interoperability organization) have established Basic Profile 1.0 based on XML, UDDI, WSDL and SOAP for web services interoperability and developed usage scenario Profile to apply Web Services in practice. In this paper, to verify suitability Web Services interoperability between heterogeneous two applications, have design and implements the Book Information Web Services that based on the Web Services Client of J2SE platform and the Web Services Server of .NET platform, so that analysis and verify the service by adaptation of WS-I Basic Profile.
Due to the development of internet technology and the rapid increase of internet data, various studies are actively conducted on how to use and analyze internet data for various purposes. In particular, in recent years, a number of studies have been performed on the applications of text mining techniques in order to overcome the limitations of the current application of structured data. Especially, there are various studies on sentimental analysis to score opinions based on the distribution of polarity such as positivity or negativity of vocabularies or sentences of the texts in documents. As a part of such studies, this study tries to predict ups and downs of stock prices of companies by performing sentimental analysis on news contexts of the particular companies in the Internet. A variety of news on companies is produced online by different economic agents, and it is diffused quickly and accessed easily in the Internet. So, based on inefficient market hypothesis, we can expect that news information of an individual company can be used to predict the fluctuations of stock prices of the company if we apply proper data analysis techniques. However, as the areas of corporate management activity are different, an analysis considering characteristics of each company is required in the analysis of text data based on machine-learning. In addition, since the news including positive or negative information on certain companies have various impacts on other companies or industry fields, an analysis for the prediction of the stock price of each company is necessary. Therefore, this study attempted to predict changes in the stock prices of the individual companies that applied a sentimental analysis of the online news data. Accordingly, this study chose top company in KOSPI 200 as the subjects of the analysis, and collected and analyzed online news data by each company produced for two years on a representative domestic search portal service, Naver. In addition, considering the differences in the meanings of vocabularies for each of the certain economic subjects, it aims to improve performance by building up a lexicon for each individual company and applying that to an analysis. As a result of the analysis, the accuracy of the prediction by each company are different, and the prediction accurate rate turned out to be 56% on average. Comparing the accuracy of the prediction of stock prices on industry sectors, 'energy/chemical', 'consumer goods for living' and 'consumer discretionary' showed a relatively higher accuracy of the prediction of stock prices than other industries, while it was found that the sectors such as 'information technology' and 'shipbuilding/transportation' industry had lower accuracy of prediction. The number of the representative companies in each industry collected was five each, so it is somewhat difficult to generalize, but it could be confirmed that there was a difference in the accuracy of the prediction of stock prices depending on industry sectors. In addition, at the individual company level, the companies such as 'Kangwon Land', 'KT & G' and 'SK Innovation' showed a relatively higher prediction accuracy as compared to other companies, while it showed that the companies such as 'Young Poong', 'LG', 'Samsung Life Insurance', and 'Doosan' had a low prediction accuracy of less than 50%. In this paper, we performed an analysis of the share price performance relative to the prediction of individual companies through the vocabulary of pre-built company to take advantage of the online news information. In this paper, we aim to improve performance of the stock prices prediction, applying online news information, through the stock price prediction of individual companies. Based on this, in the future, it will be possible to find ways to increase the stock price prediction accuracy by complementing the problem of unnecessary words that are added to the sentiment dictionary.
People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.
From January 2020 to October 2021, more than 500,000 academic studies related to COVID-19 (Coronavirus-2, a fatal respiratory syndrome) have been published. The rapid increase in the number of papers related to COVID-19 is putting time and technical constraints on healthcare professionals and policy makers to quickly find important research. Therefore, in this study, we propose a method of extracting useful information from text data of extensive literature using LDA and Word2vec algorithm. Papers related to keywords to be searched were extracted from papers related to COVID-19, and detailed topics were identified. The data used the CORD-19 data set on Kaggle, a free academic resource prepared by major research groups and the White House to respond to the COVID-19 pandemic, updated weekly. The research methods are divided into two main categories. First, 41,062 articles were collected through data filtering and pre-processing of the abstracts of 47,110 academic papers including full text. For this purpose, the number of publications related to COVID-19 by year was analyzed through exploratory data analysis using a Python program, and the top 10 journals under active research were identified. LDA and Word2vec algorithm were used to derive research topics related to COVID-19, and after analyzing related words, similarity was measured. Second, papers containing 'vaccine' and 'treatment' were extracted from among the topics derived from all papers, and a total of 4,555 papers related to 'vaccine' and 5,971 papers related to 'treatment' were extracted. did For each collected paper, detailed topics were analyzed using LDA and Word2vec algorithms, and a clustering method through PCA dimension reduction was applied to visualize groups of papers with similar themes using the t-SNE algorithm. A noteworthy point from the results of this study is that the topics that were not derived from the topics derived for all papers being researched in relation to COVID-19 (