• Title/Summary/Keyword: GREEDY

Search Result 428, Processing Time 0.024 seconds

Self-optimizing feature selection algorithm for enhancing campaign effectiveness (캠페인 효과 제고를 위한 자기 최적화 변수 선택 알고리즘)

  • Seo, Jeoung-soo;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.173-198
    • /
    • 2020
  • For a long time, many studies have been conducted on predicting the success of campaigns for customers in academia, and prediction models applying various techniques are still being studied. Recently, as campaign channels have been expanded in various ways due to the rapid revitalization of online, various types of campaigns are being carried out by companies at a level that cannot be compared to the past. However, customers tend to perceive it as spam as the fatigue of campaigns due to duplicate exposure increases. Also, from a corporate standpoint, there is a problem that the effectiveness of the campaign itself is decreasing, such as increasing the cost of investing in the campaign, which leads to the low actual campaign success rate. Accordingly, various studies are ongoing to improve the effectiveness of the campaign in practice. This campaign system has the ultimate purpose to increase the success rate of various campaigns by collecting and analyzing various data related to customers and using them for campaigns. In particular, recent attempts to make various predictions related to the response of campaigns using machine learning have been made. It is very important to select appropriate features due to the various features of campaign data. If all of the input data are used in the process of classifying a large amount of data, it takes a lot of learning time as the classification class expands, so the minimum input data set must be extracted and used from the entire data. In addition, when a trained model is generated by using too many features, prediction accuracy may be degraded due to overfitting or correlation between features. Therefore, in order to improve accuracy, a feature selection technique that removes features close to noise should be applied, and feature selection is a necessary process in order to analyze a high-dimensional data set. Among the greedy algorithms, SFS (Sequential Forward Selection), SBS (Sequential Backward Selection), SFFS (Sequential Floating Forward Selection), etc. are widely used as traditional feature selection techniques. It is also true that if there are many risks and many features, there is a limitation in that the performance for classification prediction is poor and it takes a lot of learning time. Therefore, in this study, we propose an improved feature selection algorithm to enhance the effectiveness of the existing campaign. The purpose of this study is to improve the existing SFFS sequential method in the process of searching for feature subsets that are the basis for improving machine learning model performance using statistical characteristics of the data to be processed in the campaign system. Through this, features that have a lot of influence on performance are first derived, features that have a negative effect are removed, and then the sequential method is applied to increase the efficiency for search performance and to apply an improved algorithm to enable generalized prediction. Through this, it was confirmed that the proposed model showed better search and prediction performance than the traditional greed algorithm. Compared with the original data set, greed algorithm, genetic algorithm (GA), and recursive feature elimination (RFE), the campaign success prediction was higher. In addition, when performing campaign success prediction, the improved feature selection algorithm was found to be helpful in analyzing and interpreting the prediction results by providing the importance of the derived features. This is important features such as age, customer rating, and sales, which were previously known statistically. Unlike the previous campaign planners, features such as the combined product name, average 3-month data consumption rate, and the last 3-month wireless data usage were unexpectedly selected as important features for the campaign response, which they rarely used to select campaign targets. It was confirmed that base attributes can also be very important features depending on the type of campaign. Through this, it is possible to analyze and understand the important characteristics of each campaign type.

Query Slipping Prevention for Trajectory-based Contents Publishing and Subscribing in Wireless Sensor Networks (무선 센서 네트워크에서의 궤도 기반 콘텐츠 발간 및 구독을 위한 질의 이탈 방지)

  • Tscha, Yeong-Hwan
    • Journal of KIISE:Information Networking
    • /
    • v.32 no.4
    • /
    • pp.525-534
    • /
    • 2005
  • This paper is concerned with the query slipping and its prevention for trajectory-based matchmaking service in wireless sensor networks. The problem happens when a query propagating along a subscribe trajectory moves through a publish trajectory without obtaining desired information, even though two trajectories intersect geometrically. There follows resubmission of the query or initiation of another subscribe trajectory Thus, query slipping results in considerable time delay and in the worst, looping in the trajectory or query flooding the network. We address the problem formally and suggest a solution. First, the area where nodes are distributed is logically partitioned into smaller grids, and a grid-based multicast next-hop selection algorithm is proposed. Our algorithm not only attempts to make the trajectory straight but also considers the nodal density of recipient nodes and the seamless grid-by-grid multicast. We prove that the publishing and subscribing using the algorithm eventually eliminate the possibility of the slipping. It toms out that our algorithm dissipates significantly less power of neighbor nodes, compared to the non grid-based method, as greedy forwarding, and the fixed- sized grid approach, as GAF (Geographical Adaptive Fidelity)

Decision Tree based Disambiguation of Semantic Roles for Korean Adverbial Postpositions in Korean-English Machine Translation (한영 기계번역에서 결정 트리 학습에 의한 한국어 부사격 조사의 의미 중의성 해소)

  • Park, Seong-Bae;Zhang, Byoung-Tak;Kim, Yung-Taek
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.6
    • /
    • pp.668-677
    • /
    • 2000
  • Korean has the characteristics that case postpositions determine the syntactic roles of phrases and a postposition may have more than one meanings. In particular, the adverbial postpositions make translation from Korean to English difficult, because they can have various meanings. In this paper, we describe a method for resolving such semantic ambiguities of Korean adverbial postpositions using decision trees. The training examples for decision tree induction are extracted from a corpus consisting of 0.5 million words, and the semantic roles for adverbial postpositions are classified into 25 classes. The lack of training examples in decision tree induction is overcome by clustering words into classes using a greedy clustering algorithm. The cross validation results show that the presented method achieved 76.2% of precision on the average, which means 26.0% improvement over the method determining the semantic role of an adverbial postposition as the most frequently appearing role.

  • PDF

Location Strategy of Sports Oulets to Maximize the Market Share (시장 점유율을 최대로 할 수 있는 스포츠용품점 위치 결정 전략)

  • Lee, Sang-Un;Lee, Young-Sook;Choi, Seong-Beom;Han, Tae-Yong
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.13 no.3
    • /
    • pp.93-101
    • /
    • 2013
  • This paper suggests optimal location algorithm of new firm $A(F_A)^{\prime}s$ p(p$B(F_B)$ already operating q outlets of sports in the market. This algorithm selects top q nodes among $V=V{\backslash}F_B$ nodes that covers maximum nodes based on the shortest distance. Then, q nodes choose next node that has a maximum cover with inclusion-exclusion principle. At the time of same number of cardinality in q sets to pre-defined q, we select the maximum cover node set. This algorithm called by competitive algorithm. The competitive algorithm simply decides the optimal location of the outlets p=1,2,3,4 for q=5. Also, we show that the market share of competitive algorithm can be maximize.

Sparse Signal Recovery with Parallel Orthogonal Matching Pursuit for Multiple Measurement Vectors (병렬OMP 기법을 통한 복수 측정 벡터기반 성긴 신호의 복원)

  • Park, Jeonghong;Ban, Tae Won;Jung, Bang Chul
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.17 no.10
    • /
    • pp.2252-2258
    • /
    • 2013
  • In this paper, parallel orthogonal matching pursuit (POMP) is proposed to supplement the simultaneous orthogonal matching pursuit (S-OMP) which has been widely used as a greedy algorithm for sparse signal recovery for multiple measurement vector (MMV) problem. The process of POMP is simple but effective: (1) multiple indexes maximally correlated with the observation vector are chosen at the first iteration, (2) the conventional S-OMP process is carried out in parallel for each selected index, (3) the index set which yields the minimum residual is selected for reconstructing the original sparse signal. Empirical simulations show that POMP for MMV outperforms than the conventional S-OMP both in terms of exact recovery ratio (ERR) and mean-squared error (MSE).

Hyper-Rectangle Based Prototype Selection Algorithm Preserving Class Regions (클래스 영역을 보존하는 초월 사각형에 의한 프로토타입 선택 알고리즘)

  • Baek, Byunghyun;Euh, Seongyul;Hwang, Doosung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.3
    • /
    • pp.83-90
    • /
    • 2020
  • Prototype selection offers the advantage of ensuring low learning time and storage space by selecting the minimum data representative of in-class partitions from the training data. This paper designs a new training data generation method using hyper-rectangles that can be applied to general classification algorithms. Hyper-rectangular regions do not contain different class data and divide the same class space. The median value of the data within a hyper-rectangle is selected as a prototype to form new training data, and the size of the hyper-rectangle is adjusted to reflect the data distribution in the class area. A set cover optimization algorithm is proposed to select the minimum prototype set that represents the whole training data. The proposed method reduces the time complexity that requires the polynomial time of the set cover optimization algorithm by using the greedy algorithm and the distance equation without multiplication. In experimented comparison with hyper-sphere prototype selections, the proposed method is superior in terms of prototype rate and generalization performance.

Performance Improvement of Cooperating Agents through Balance between Intensification and Diversification (강화와 다양화의 조화를 통한 협력 에이전트 성능 개선에 관한 연구)

  • 이승관;정태충
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.40 no.6
    • /
    • pp.87-94
    • /
    • 2003
  • One of the important fields for heuristic algorithm is how to balance between Intensification and Diversification. Ant Colony Optimization(ACO) is a new meta heuristic algorithm to solve hard combinatorial optimization problem. It is a population based approach that uses exploitation of positive feedback as well as Breedy search It was first Proposed for tackling the well known Traveling Salesman Problem(TSP). In this paper, we deal with the performance improvement techniques through balance the Intensification and Diversification in Ant Colony System(ACS). First State Transition considering the number of times that agents visit about each edge makes agents search more variously and widen search area. After setting up criteria which divide elite tour that receive Positive Intensification about each tour, we propose a method to do addition Intensification by the criteria. Implemetation of the algorithm to solve TSP and the performance results under various conditions are conducted, and the comparision between the original An and the proposed method is shown. It turns out that our proposed method can compete with the original ACS in terms of solution quality and computation speed to these problem.

QoS-, Energy- and Cost-efficient Resource Allocation for Cloud-based Interactive TV Applications

  • Kulupana, Gosala;Talagala, Dumidu S.;Arachchi, Hemantha Kodikara;Fernando, Anil
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.6 no.3
    • /
    • pp.158-167
    • /
    • 2017
  • Internet-based social and interactive video applications have become major constituents of the envisaged applications for next-generation multimedia networks. However, inherently dynamic network conditions, together with varying user expectations, pose many challenges for resource allocation mechanisms for such applications. Yet, in addition to addressing these challenges, service providers must also consider how to mitigate their operational costs (e.g., energy costs, equipment costs) while satisfying the end-user quality of service (QoS) expectations. This paper proposes a heuristic solution to the problem, where the energy incurred by the applications, and the monetary costs associated with the service infrastructure, are minimized while simultaneously maximizing the average end-user QoS. We evaluate the performance of the proposed solution in terms of serving probability, i.e., the likelihood of being able to allocate resources to groups of users, the computation time of the resource allocation process, and the adaptability and sensitivity to dynamic network conditions. The proposed method demonstrates improvements in serving probability of up to 27%, in comparison with greedy resource allocation schemes, and a several-orders-of-magnitude reduction in computation time, compared to the linear programming approach, which significantly reduces the service-interrupted user percentage when operating under variable network conditions.

Mobility-Aware Service Migration (MASM) Algorithms for Multi-Access Edge Computing (멀티 액세스 엣지 컴퓨팅을 위한 Mobility-Aware Service Migration (MASM) 알고리즘)

  • Hamzah, Haziq;Le, Duc-Tai;Kim, Moonseong;Choo, Hyunseung
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.1-8
    • /
    • 2020
  • In order to reach Ultra-Reliable Low-Latency communication, one of 5G aims, Multi-access Edge Computing paradigm was born. The idea of this paradigm is to bring cloud computing technologies closer to the network edge. User services are hosted in multiple Edge Clouds, deployed at the edge of the network distributedly, to reduce the service latency. For mobile users, migrating their services to the most proper Edge Clouds for maintaining a Quality of Service is a non-convex problem. The service migration problem becomes more complex in high mobility scenarios. The goal of the study is to observe how user mobility affects the selection of Edge Cloud during a fixed mobility path. Mobility-Aware Service Migration (MASM) is proposed to optimize service migration based on two main parameters: routing cost and service migration cost, during a high mobility scenario. The performance of the proposed algorithm is compared with an existing greedy algorithm.

Optimal Decisions on the Quantity and Locations of Ambulances for the Timely Response to Emergency Requests (출동 응답 향상을 위한 적정 구급차 수량 및 배치 위치 결정 연구)

  • Jeong, Yonghun;Jeong, Heena;Ko, Jeonghan
    • Fire Science and Engineering
    • /
    • v.31 no.3
    • /
    • pp.137-143
    • /
    • 2017
  • A sufficient number of ambulances are critical for preventing delayed vehicle dispatch for emergency patients. This study presents effective methodologies for evaluating the effects of ambulance quantities on availability. The statistical properties of the emergency requests and responses were analyzed for a city in Korea. The inter-request times were modeled by statistical distributions. The ambulance dispatch was modeled using simulation, reflecting the shared dispatch among the city districts. The simulation results revealed that the existing ambulance quantity could successfully meet the majority of the requests, but more vehicles were desirable for improvement. The locations of the additional vehicles were determined efficiently by simulations with a greedy approach. The simulations with added vehicles showed a significantly better coverage of the emergency calls. This research can help design improved emergency vehicle operations, and help save lives.