• Title/Summary/Keyword: Internet models

Search Result 1,391, Processing Time 0.023 seconds

Prediction Model of Real Estate ROI with the LSTM Model based on AI and Bigdata

  • Lee, Jeong-hyun;Kim, Hoo-bin;Shim, Gyo-eon
    • International journal of advanced smart convergence
    • /
    • v.11 no.1
    • /
    • pp.19-27
    • /
    • 2022
  • Across the world, 'housing' comprises a significant portion of wealth and assets. For this reason, fluctuations in real estate prices are highly sensitive issues to individual households. In Korea, housing prices have steadily increased over the years, and thus many Koreans view the real estate market as an effective channel for their investments. However, if one purchases a real estate property for the purpose of investing, then there are several risks involved when prices begin to fluctuate. The purpose of this study is to design a real estate price 'return rate' prediction model to help mitigate the risks involved with real estate investments and promote reasonable real estate purchases. Various approaches are explored to develop a model capable of predicting real estate prices based on an understanding of the immovability of the real estate market. This study employs the LSTM method, which is based on artificial intelligence and deep learning, to predict real estate prices and validate the model. LSTM networks are based on recurrent neural networks (RNN) but add cell states (which act as a type of conveyer belt) to the hidden states. LSTM networks are able to obtain cell states and hidden states in a recursive manner. Data on the actual trading prices of apartments in autonomous districts between January 2006 and December 2019 are collected from the Actual Trading Price Disclosure System of the Ministry of Land, Infrastructure and Transport (MOLIT). Additionally, basic data on apartments and commercial buildings are collected from the Public Data Portal and Seoul Metropolitan Government's data portal. The collected actual trading price data are scaled to monthly average trading amounts, and each data entry is pre-processed according to address to produce 168 data entries. An LSTM model for return rate prediction is prepared based on a time series dataset where the training period is set as April 2015~August 2017 (29 months), the validation period is set as September 2017~September 2018 (13 months), and the test period is set as December 2018~December 2019 (13 months). The results of the return rate prediction study are as follows. First, the model achieved a prediction similarity level of almost 76%. After collecting time series data and preparing the final prediction model, it was confirmed that 76% of models could be achieved. All in all, the results demonstrate the reliability of the LSTM-based model for return rate prediction.

A Study on the Media Recommendation System with Time Period Considering the Consumer Contextual Information Using Public Data (공공 데이터 기반 소비자 상황을 고려한 시간대별 미디어 추천 시스템 연구)

  • Kim, Eunbi;Li, Qinglong;Chang, Pilsik;Kim, Jaekyeong
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.4
    • /
    • pp.95-117
    • /
    • 2022
  • With the emergence of various media types due to the development of Internet technology, advertisers have difficulty choosing media suitable for corporate advertising strategies. There are challenging to effectively reflect consumer contextual information when advertising media is selected based on traditional marketing strategies. Thus, a recommender system is needed to analyze consumers' past data and provide advertisers with personalized media based on the information consumers needs. Since the traditional recommender system provides recommendation services based on quantitative preference information, there is difficult to reflect various contextual information. This study proposes a methodology that uses deep learning to recommend personalized media to advertisers using consumer contextual information such as consumers' media viewing time, residence area, age, and gender. This study builds a recommender system using media & consumer research data provided by the Korea Broadcasting Advertising Promotion Corporation. Additionally, we evaluate the recommendation performance compared with several benchmark models. As a result of the experiment, we confirmed that the recommendation model reflecting the consumer's contextual information showed higher accuracy than the benchmark model. We expect to contribute to helping advertisers make effective decisions when selecting customized media based on various contextual information of consumers.

Big Data Management in Structured Storage Based on Fintech Models for IoMT using Machine Learning Techniques (기계학습법을 이용한 IoMT 핀테크 모델을 기반으로 한 구조화 스토리지에서의 빅데이터 관리 연구)

  • Kim, Kyung-Sil
    • Advanced Industrial SCIence
    • /
    • v.1 no.1
    • /
    • pp.7-15
    • /
    • 2022
  • To adopt the development in the medical scenario IoT developed towards the advancement with the processing of a large amount of medical data defined as an Internet of Medical Things (IoMT). The vast range of collected medical data is stored in the cloud in the structured manner to process the collected healthcare data. However, it is difficult to handle the huge volume of the healthcare data so it is necessary to develop an appropriate scheme for the healthcare structured data. In this paper, a machine learning mode for processing the structured heath care data collected from the IoMT is suggested. To process the vast range of healthcare data, this paper proposed an MTGPLSTM model for the processing of the medical data. The proposed model integrates the linear regression model for the processing of healthcare information. With the developed model outlier model is implemented based on the FinTech model for the evaluation and prediction of the COVID-19 healthcare dataset collected from the IoMT. The proposed MTGPLSTM model comprises of the regression model to predict and evaluate the planning scheme for the prevention of the infection spreading. The developed model performance is evaluated based on the consideration of the different classifiers such as LR, SVR, RFR, LSTM and the proposed MTGPLSTM model and the different size of data as 1GB, 2GB and 3GB is mainly concerned. The comparative analysis expressed that the proposed MTGPLSTM model achieves ~4% reduced MAPE and RMSE value for the worldwide data; in case of china minimal MAPE value of 0.97 is achieved which is ~ 6% minimal than the existing classifier leads.

Preprocessing Technique for Malicious Comments Detection Considering the Form of Comments Used in the Online Community (온라인 커뮤니티에서 사용되는 댓글의 형태를 고려한 악플 탐지를 위한 전처리 기법)

  • Kim Hae Soo;Kim Mi Hui
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.12 no.3
    • /
    • pp.103-110
    • /
    • 2023
  • With the spread of the Internet, anonymous communities emerged along with the activation of communities for communication between people, and many users are doing harm to others, such as posting aggressive posts and leaving comments using anonymity. In the past, administrators directly checked posts and comments, then deleted and blocked them, but as the number of community users increased, they reached a level that managers could not continue to monitor. Initially, word filtering techniques were used to prevent malicious writing from being posted in a form that could not post or comment if a specific word was included, but they avoided filtering in a bypassed form, such as using similar words. As a way to solve this problem, deep learning was used to monitor posts posted by users in real-time, but recently, the community uses words that can only be understood by the community or from a human perspective, not from a general Korean word. There are various types and forms of characters, making it difficult to learn everything in the artificial intelligence model. Therefore, in this paper, we proposes a preprocessing technique in which each character of a sentence is imaged using a CNN model that learns the consonants, vowel and spacing images of Korean word and converts characters that can only be understood from a human perspective into characters predicted by the CNN model. As a result of the experiment, it was confirmed that the performance of the LSTM, BiLSTM and CNN-BiLSTM models increased by 3.2%, 3.3%, and 4.88%, respectively, through the proposed preprocessing technique.

A study on the detection of fake news - The Comparison of detection performance according to the use of social engagement networks (그래프 임베딩을 활용한 코로나19 가짜뉴스 탐지 연구 - 사회적 참여 네트워크의 이용 여부에 따른 탐지 성능 비교)

  • Jeong, Iitae;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.197-216
    • /
    • 2022
  • With the development of Internet and mobile technology and the spread of social media, a large amount of information is being generated and distributed online. Some of them are useful information for the public, but others are misleading information. The misleading information, so-called 'fake news', has been causing great harm to our society in recent years. Since the global spread of COVID-19 in 2020, much of fake news has been distributed online. Unlike other fake news, fake news related to COVID-19 can threaten people's health and even their lives. Therefore, intelligent technology that automatically detects and prevents fake news related to COVID-19 is a meaningful research topic to improve social health. Fake news related to COVID-19 has spread rapidly through social media, however, there have been few studies in Korea that proposed intelligent fake news detection using the information about how the fake news spreads through social media. Under this background, we propose a novel model that uses Graph2vec, one of the graph embedding methods, to effectively detect fake news related to COVID-19. The mainstream approaches of fake news detection have focused on news content, i.e., characteristics of the text, but the proposed model in this study can exploit information transmission relationships in social engagement networks when detecting fake news related to COVID-19. Experiments using a real-world data set have shown that our proposed model outperforms traditional models from the perspectives of prediction accuracy.

The Effects of Non Verbal Communication of Restaurant Employees on Customer Emotion, Customer Satisfaction, Customer Trust, and Revisit Intention (외식업 직원의 비언어적 커뮤니케이션이 고객감정, 고객만족, 고객신뢰 그리고 재방문의도에 미치는 영향)

  • Kim, Bo-Yeong;Jun, Jae-Hyeon;Han, Sang-Ho
    • The Korean Journal of Franchise Management
    • /
    • v.9 no.3
    • /
    • pp.45-55
    • /
    • 2018
  • Purpose - Non-verbal Communication with customers in restaurant business can play an important role because it affects customer behavior and attitudes as a means to develop and maintain long-term relationships with customers. The purpose of this study is to analyze the effect of non-verbal communication with customers and the effect of the influence on customer satisfaction, trust, and revisit intention. Research design, data, methodology - In order to verify the research models and hypotheses of this study, questions were prepared for each variable and data were collected through questionnaires. The questionnaire survey was conducted from March 27, 2018 to April 17, 2018, for those who agreed with the citizens of the Jeju area who visited the restaurant recently. 50 out of 100 were conducted by internet survey and 50 were surveyed. Thus, a total of 100 responses were used using structural equation modeling with Smartpls 3.0. Results - The results of the study are as follows. First, non-verbal communication has a significant impact on customer emotion. Second customer emotion have a significant impact on customer trust and satisfaction. Third, Customer satisfaction had positive a significant effect on revisit intention. Fourth, Customer trust had positive a significant effect on revisit intention. Conclusions - The implications of this study are following as: The food service company should continuously provide non-verbal communication training to employees so that they can respond to customers with the right attitude and bright smile. In particular, in the case of restaurant franchises, customer response manuals should be created and distributed to the franchisees, and a regular training program for the franchisees should be implemented to provide the same service to the customer. Second, CEOs should have to worry about what kind of experience he or she has left since leaving the store. It is also necessary to constantly look at what customers experience in their stores or in their brands, and what emotions they form through their experiences. Third, the more satisfied or trusted customers are formed through the service of the employee, the more loyal the restaurant business will be, and the more likely it is to make continuous revisit and positive word-of-mouth activities..

Environmental Equity Analysis of Fine Dust in Daegu Using MGWR and KT Sensor Data (다중 스케일 지리가중회귀 모형과 KT 측정기 자료를 활용한 대구시 미세먼지에 대한 환경적 형평성 분석)

  • Euna CHO;Byong-Woon JUN
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.26 no.4
    • /
    • pp.218-236
    • /
    • 2023
  • This study attempted to analyze the environmental equity of fine dust(PM10) in Daegu using MGWR(Multi-scale Geographically Weighted Regression) and KT(Korea Telecom Corporation) sensor data. Existing national monitoring network data for measuring fine dust are collected at a small number of ground-based stations that are sparsely distributed in a large area. To complement these drawbacks, KT sensor data with a large number of IoT(Internet of Things) stations densely distributed were used in this study. The MGWR model was used to deal with spatial heterogeneity and multi-scale contextual effects in the spatial relationships between fine dust concentration and socioeconomic variables. Results indicate that there existed an environmental inequity by land value and foreigner ratio in the spatial distribution of fine dust in Daegu metropolitan city. Also, the MGWR model showed better the explanatory power than Ordinary Least Square(OLS) and Geographically Weighted Regression(GWR) models in explaining the spatial relationships between the concentration of fine dust and socioeconomic variables. This study demonstrated the potential of KT sensor data as a supplement to the existing national monitoring network data for measuring fine dust.

The Effect of the Introduction Characteristics of Cloud Computing Services on the Performance Expectancy of Firms: Setting Up Innovativeness as the Moderator (클라우드 컴퓨팅 서비스의 도입특성이 기업의 인지된 기대성과에 미치는 영향: 기업의 혁신채택성향을 조절변수로)

  • Jae Su Lim;Jay In Oh
    • Information Systems Review
    • /
    • v.19 no.1
    • /
    • pp.75-100
    • /
    • 2017
  • Today, firms are constantly transforming and innovating to survive under the rapidly changing business environment. The introduction of cloud computing services has become popular throughout society as a whole and is expected to result in many changes and developments not only in firms and but also in the public sector subject to innovation. The purpose of this study is to investigate the effect of the characteristics of cloud computing services on the perceived expected performance according to innovativeness based on innovation diffusion theory. The results of the analysis of the data collected from this research are as follows. The convenience and understanding of individuals' work as well as the benefits of cloud computing services to them depend on the innovative trend of cloud computing services. Further, the expectations for personal benefit and those for organizational benefit of cloud computing services are different from each other. Leading firms in the global market have been actively engaged in the utilization of cloud computing services in the public sector as well as in private firms. In consideration of the importance of cloud computing services, using cloud computing services as the target of innovation diffusion research is important. The results of the study are expected to contribute to developing future research models for the diffusion of new technologies, such as big data, digital convergence, and Internet of Things.

Building an Efficient Supply Chain by reduction of lead time with a Focus on Korea Server Manufacturer (리드타임 감소에 의한 효율적 공급체인 구축 - 국내 서버 공급체인을 대상으로 -)

  • 신용석;김태현;문성암
    • Journal of Distribution Research
    • /
    • v.6 no.2
    • /
    • pp.1-17
    • /
    • 2002
  • The recent dot-com craze has been one of the main causes that accelerated the growth of internet-related companies in diversity as well as in size. Meanwhile, the domestic market of supplies and equipment for internet businesses has been dominated by major foreign companies. To regain their market positions, the domestic manufacturers had to find the way to build up their competitive advantages, such as meeting their customers needs and reducing overall costs. In this study, one domestic PC server manufacturer, which competes fiercely with foreign manufacturers for the top place, has been chosen as a model to evaluate its current supply chain and to find an area that can be improved for a better performance. System Dynamics is used throughout the study. The central concept to system dynamics is understanding how all the objects in a system interact with one another. It focuses on feedback and secondary effects to think through how a strategy might or might not work, depending on how organizational changes are received, and what kinds of consequences emerge. Then, computerized models were built for simulations, each with different conditions, and, finally, the results were evaluated based on some criteria which are considered to be important and meaningful. The inefficiency that exists in the supply chain was proved to be a thirty-day long purchasing order leadtime, and it was expected that more effective supply chain could be formed if the leadtme were reduced to 14 days or 7 days. The results of simulations showed that the overall expected costs in supply chain was the least with the purchasing leadtime being 7 days. The lower average number of parts held as inventory, along with the reduced lost sales, acted as the factor reducing the expected overall costs. Although there was a slight increase in the average number of final products held as inventory and the total ordering cost, the benefits from lower parts inventory and reduced lost sales were large enough to justify the overall cost reduction.

  • PDF

Improving the Accuracy of Document Classification by Learning Heterogeneity (이질성 학습을 통한 문서 분류의 정확성 향상 기법)

  • Wong, William Xiu Shun;Hyun, Yoonjin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.21-44
    • /
    • 2018
  • In recent years, the rapid development of internet technology and the popularization of smart devices have resulted in massive amounts of text data. Those text data were produced and distributed through various media platforms such as World Wide Web, Internet news feeds, microblog, and social media. However, this enormous amount of easily obtained information is lack of organization. Therefore, this problem has raised the interest of many researchers in order to manage this huge amount of information. Further, this problem also required professionals that are capable of classifying relevant information and hence text classification is introduced. Text classification is a challenging task in modern data analysis, which it needs to assign a text document into one or more predefined categories or classes. In text classification field, there are different kinds of techniques available such as K-Nearest Neighbor, Naïve Bayes Algorithm, Support Vector Machine, Decision Tree, and Artificial Neural Network. However, while dealing with huge amount of text data, model performance and accuracy becomes a challenge. According to the type of words used in the corpus and type of features created for classification, the performance of a text classification model can be varied. Most of the attempts are been made based on proposing a new algorithm or modifying an existing algorithm. This kind of research can be said already reached their certain limitations for further improvements. In this study, aside from proposing a new algorithm or modifying the algorithm, we focus on searching a way to modify the use of data. It is widely known that classifier performance is influenced by the quality of training data upon which this classifier is built. The real world datasets in most of the time contain noise, or in other words noisy data, these can actually affect the decision made by the classifiers built from these data. In this study, we consider that the data from different domains, which is heterogeneous data might have the characteristics of noise which can be utilized in the classification process. In order to build the classifier, machine learning algorithm is performed based on the assumption that the characteristics of training data and target data are the same or very similar to each other. However, in the case of unstructured data such as text, the features are determined according to the vocabularies included in the document. If the viewpoints of the learning data and target data are different, the features may be appearing different between these two data. In this study, we attempt to improve the classification accuracy by strengthening the robustness of the document classifier through artificially injecting the noise into the process of constructing the document classifier. With data coming from various kind of sources, these data are likely formatted differently. These cause difficulties for traditional machine learning algorithms because they are not developed to recognize different type of data representation at one time and to put them together in same generalization. Therefore, in order to utilize heterogeneous data in the learning process of document classifier, we apply semi-supervised learning in our study. However, unlabeled data might have the possibility to degrade the performance of the document classifier. Therefore, we further proposed a method called Rule Selection-Based Ensemble Semi-Supervised Learning Algorithm (RSESLA) to select only the documents that contributing to the accuracy improvement of the classifier. RSESLA creates multiple views by manipulating the features using different types of classification models and different types of heterogeneous data. The most confident classification rules will be selected and applied for the final decision making. In this paper, three different types of real-world data sources were used, which are news, twitter and blogs.