Designing an Intelligent Advertising Business Model in Seoul's Metro Network (서울지하철의 지능형 광고 비즈니스모델 설계)
-
- Journal of Intelligence and Information Systems
- /
- v.23 no.4
- /
- pp.1-31
- /
- 2017
Modern businesses are adopting new technologies to serve their markets better as well as to improve efficiency and productivity. The advertising industry has continuously experienced disruptions from the traditional channels (radio, television and print media) to new complex ones including internet, social media and mobile-based advertising. This case study focuses on proposing intelligent advertising business model in Seoul's metro network. Seoul has one of the world's busiest metro network and transports a huge number of travelers on a daily basis. The high number of travelers coupled with a well-planned metro network creates a platform where marketers can initiate engagement and interact with both customers and potential customers. In the current advertising model, advertising is on illuminated and framed posters in the stations and in-car, non-illuminated posters, and digital screens that show scheduled arrivals and departures of metros. Some stations have digital screens that show adverts but they do not have location capability. Most of the current advertising media have one key limitation: space. For posters whether illuminated or not, one space can host only one advert at a time. Empirical literatures show that there is room for improving this advertising model and eliminate the space limitation by replacing the poster adverts with digital advertising platform. This new model will not only be digital, but will also provide intelligent advertising platform that is driven by data. The digital platform will incorporate location sensing, e-commerce, and mobile platform to create new value to all stakeholders. Travel cards used in the metro will be registered and the card scanners will have a capability to capture traveler's data when travelers tap their cards. This data once analyzed will make it possible to identify different customer groups. Advertisers and marketers will then be able to target specific customer groups, customize adverts based on the targeted consumer group, and offer a wide variety of advertising formats. Format includes video, cinemagraphs, moving pictures, and animation. Different advert formats create different emotions in the customer's mind and the goal should be to use format or combination of formats that arouse the expected emotion and lead to an engagement. Combination of different formats will be more effective and this can only work in a digital platform. Adverts will be location based, ensuring that adverts will show more frequently when the metro is near the premises of an advertiser. The advertising platform will automatically detect the next station and screens inside the metro will prioritize adverts in the station where the metro will be stopping. In the mobile platform, customers who opt to receive notifications will receive them when they approach the business premises of advertiser. The mobile platform will have indoor navigation for the underground shopping malls that will allow customers to search for facilities within the mall, products they may want to buy as well as deals going on in the underground mall. To create an end-to-end solution, the mobile solution will have a capability to allow customers purchase products through their phones, get coupons for deals, and review products and shops where they have bought a product. The indoor navigation will host intelligent mobile-based advertisement and a recommendation system. The indoor navigation will have adverts such that when a customer is searching for information, the recommendation system shows adverts that are near the place traveler is searching or in the direction that the traveler is moving. These adverts will be linked to the e-commerce platform such that if a customer clicks on an advert, it leads them to the product description page. The whole system will have multi-language as well as text-to-speech capability such that both locals and tourists have no language barrier. The implications of implementing this model are varied including support for small and medium businesses operating in the underground malls, improved customer experience, new job opportunities, additional revenue to business model operator, and flexibility in advertising. The new value created will benefit all the stakeholders.
As the internet technologies develop, the geographic information system environment is changing to the web-based service. Since geospatial information of the existing Web-GIS services were developed independently, there is no interoperability to support diverse map formats. In spite of the same geospatial information object it can be used for various proposes that is duplicated in GIS separately. It needs intelligent strategies for optimal replica selection, which is identification of replication geospatial information objects. And for management of replication objects, OMG, GLOBE and GRID computing suggested related frameworks. But these researches are not thorough going enough in case of geospatial information object. This paper presents a model of location service, which is supported for optimal selection among replication and management of replication objects. It is consist of tree main services. The first is binding service which can save names and properties of object defined by users according to service offers and enable clients to search them on the service of offers. The second is location service which can manage location information with contact records. And obtains performance information by the Load Sharing Facility on system independently with contact address. The third is intelligent selection service which can obtain basic/performance information from the binding service/location service and provide both faster access and better performance characteristics by rules as intelligent model based on rough sets. For the validity of location service model, this research presents the processes of location service execution with Graphic User Interface.
If doubtlessly contributes much to agriculture and rural development. The roles can be summarized as; 1. to activate rural areas and to provide more comfortable and safe rural life with equivalent services to those in urban areas, facilitating distance education, tole-medicine, remote public services, remote entertainment etc. 2. To initiate new agricultural and rural business such as e-commerce, real estate business for satellite officies, rural tourism and virtual corporation of small-scale farms. 3. To support policy-making and evaluation on optimal farm production, disaster management, effective agro-environmental resource management etc., providing tools such as GIS. 4. To improve farm management and farming technologies by efficient farm management, risk management, effective information or knowledge transfer etc., realizing competitive and sustainable farming with safe products. 5. To provide systems and tools to secure food traceability and reliability that has been an emerging issue concerning farm products since serious contamination such as BSE and chicken flu was detected. 6. To take an important and key role for industrialization of farming or lam business enterprise, combining the above roles.
Various functional foods, marketing health and functional effects, have been distributed in the market. These products, being in forms of foods, tablets, and capsules, are likely to be mistaken as drugs. In addition, non-experts may sell these as foods, or use these for therapy. Efforts for creating health food regulations or building regulatory system for improving the current status of functional foods have been made, but these have not been communicated to consumers yet. As a result, problems of circulating functional foods for therapy or adding illegal medical to such products have persisted, which has become worse by internet media. The cause of this problem can be categorized into (1) product itself and (2) its use, but in either case, one possible cause is lack of communications with consumers. Potential problems that can be caused by functional foods include illegal substances, hazardous substances, allergic reactions, considerations when administered to patients, drug interactions, ingredients with purity or concentrations too low to be detected, products with metabolic activations, health risks from over- or under-dose of vitamin and minerals, and products with alkaloids. (Journal of Health Science, 56, Supplement (2010)). The reason why side effects related to functional foods have been increasing is that under-qualified functional food companies are exaggerating the functionality for marketing purposes. KFDA has been informing consumers, through its web pages, to address the above mentioned issues related to functional foods, but there still is room for improvement, to promote proper use of functional foods and avoid drug interactions. Specifically, to address these issues, institutionalizing to collect information on approved products and their side effects, settling reevaluation systems, and standardizing preclinical tests and clinical tests are becoming urgent. Also to provide crucial information, unified database systems, seamlessly aggregating heterogeneous data in different domains, with user interfaces enabling effective one-stop search, are crucial.
Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.
Recommender system has become one of the most important technologies in e-commerce in these days. The ultimate reason to shop online, for many consumers, is to reduce the efforts for information search and purchase. Recommender system is a key technology to serve these needs. Many of the past studies about recommender systems have been devoted to developing and improving recommendation algorithms and collaborative filtering (CF) is known to be the most successful one. Despite its success, however, CF has several shortcomings such as cold-start, sparsity, gray sheep problems. In order to be able to generate recommendations, ordinary CF algorithms require evaluations or preference information directly from users. For new users who do not have any evaluations or preference information, therefore, CF cannot come up with recommendations (Cold-star problem). As the numbers of products and customers increase, the scale of the data increases exponentially and most of the data cells are empty. This sparse dataset makes computation for recommendation extremely hard (Sparsity problem). Since CF is based on the assumption that there are groups of users sharing common preferences or tastes, CF becomes inaccurate if there are many users with rare and unique tastes (Gray sheep problem). This study proposes a new algorithm that utilizes Social Network Analysis (SNA) techniques to resolve the gray sheep problem. We utilize 'degree centrality' in SNA to identify users with unique preferences (gray sheep). Degree centrality in SNA refers to the number of direct links to and from a node. In a network of users who are connected through common preferences or tastes, those with unique tastes have fewer links to other users (nodes) and they are isolated from other users. Therefore, gray sheep can be identified by calculating degree centrality of each node. We divide the dataset into two, gray sheep and others, based on the degree centrality of the users. Then, different similarity measures and recommendation methods are applied to these two datasets. More detail algorithm is as follows: Step 1: Convert the initial data which is a two-mode network (user to item) into an one-mode network (user to user). Step 2: Calculate degree centrality of each node and separate those nodes having degree centrality values lower than the pre-set threshold. The threshold value is determined by simulations such that the accuracy of CF for the remaining dataset is maximized. Step 3: Ordinary CF algorithm is applied to the remaining dataset. Step 4: Since the separated dataset consist of users with unique tastes, an ordinary CF algorithm cannot generate recommendations for them. A 'popular item' method is used to generate recommendations for these users. The F measures of the two datasets are weighted by the numbers of nodes and summed to be used as the final performance metric. In order to test performance improvement by this new algorithm, an empirical study was conducted using a publically available dataset - the MovieLens data by GroupLens research team. We used 100,000 evaluations by 943 users on 1,682 movies. The proposed algorithm was compared with an ordinary CF algorithm utilizing 'Best-N-neighbors' and 'Cosine' similarity method. The empirical results show that F measure was improved about 11% on average when the proposed algorithm was used