• 제목/요약/키워드: reward time

검색결과 166건 처리시간 0.03초

온-오프라인 채널에서 운영하는 고객보상프로그램의 보상채널과 보상시점에 따른 효과 분석 -백화점과 온라인 종합몰을 중심으로- (The Effect of Reward Channel and Reward Time of Customer Loyalty Programs for On-offline Channels -Focusing on Department Stores and Online Shopping Stores-)

  • 박민정
    • 한국의류학회지
    • /
    • 제37권4호
    • /
    • pp.467-481
    • /
    • 2013
  • The study examined the main effect of the reward channel and reward time of customer loyalty programs for on-offline shopping channels; in addition, it investigated the interaction effect of the reward channels and merchandise as well as the interaction effect of the reward time and merchandise. An online apparel shopping web experiment was conducted with a 2 (reward channel: online channel reward vs. offline channel reward) ${\times}2$ (reward time: immediate vs. delayed) ${\times}2$ (merchandise: online channel product vs. offline channel products) between-subject factorial design. An online shopping channel was considered the core-shopping channel and a department store was considered the cross-shopping channel. Loyalty program value, core-channel loyalty and cross-channel loyalty were measured as dependent variables. A total of 845 shoppers (who had experiences in shopping in both channels) participated in the experiment. The results of the study revealed (1) the main effect of the reward channel on loyalty program value, core-channel loyalty and cross-channel loyalty [online>offline channel rewards], (2) the main effect of reward time on loyalty program value, core-channel loyalty and cross-channel loyalty [immediate>delayed reward], and (3) the interaction effect of the reward channel and merchandise on loyalty program value, core-channel loyalty, and cross-channel loyalty. (4) Finally the study found that loyalty program value affected cross-channel loyalty indirectly through core-channel loyalty. This study suggested diverse theoretical and managerial implications for multi-channel retailers.

외식업 홈페이지 고객 보상 프로그램이 신뢰와 몰입 및 고객 충성도에 미치는 영향 (The Effect of the Reward Program in Foodservice Homepages on Customer Trust, Commitment and Loyalty)

  • 김옥란;김지응;최원식
    • 한국조리학회지
    • /
    • 제15권4호
    • /
    • pp.313-330
    • /
    • 2009
  • 본 연구는 외식업체의 판매 촉진 수단으로 가장 폭 넓게 활용되고 있는 온라인 웹 홈페이지에서의 고객 보상 프로그램이 고객의 신뢰와 몰입 및 고객 충성도에 미치는 영향에 대해 인과 관계를 밝히는데 목적을 갖고 연구하였는데, 그 결과를 요약하면 다음과 같다. 고객 보상 프로그램 내 보상 특성이 고객의 신뢰와 몰입에 미치는 영향을 검증한 결과 먼저 보상 시점의 경우 지연시점과 즉시시점의 보상 모두 고객의 신뢰, 몰입에 통계적으로 유의한 정(+)의 영향을 미치는 것으로 나타나, 고객 보상 프로그램의 보상 시점은 고객의 신뢰, 몰입의 향상에 중요한 요인임을 시사해 주었다. 또한 보상 유형의 경우도 고객의 신뢰, 몰입의 향상에 중요한 요인임을 알 수 있으며, 특히 간접적인 보상은 지속적인 고객의 몰입이나 충성도를 높이는데 유의적인 것으로 볼 수 있으며, 보상 속성의 경우 신뢰 형성에는 경제적 보상이 유의적인 속성임을 확인시켜 주었다.

  • PDF

인지양식 유형과 보상의 제시형태에 따른 아동의 만족지연능력 발달 (The Development of Delay of Gratification by Cognitive Style and Reward Presentation)

  • 허수경;이경님
    • 아동학회지
    • /
    • 제17권2호
    • /
    • pp.221-233
    • /
    • 1996
  • The purpose of the present study was to investigate the effects of age, sex, cognitive style and reward presentation on delay of gratification. The subjects of this study were 120 children 4, 6 and 8 years of age attending preschool and an elementary school in Pusan. They were identified as impulsive or reflective according to their performance on Kagan's Matching Familiar Figures Test. The levels of reward presentation consisted of the reward which was presented physically and the reward which wasn't presented physically. Length of waiting time was recorded as the measure of maintenance of delay of gratification. The data of this study were analyzed with Two-way ANOVA, Duncan's Multiple Range Test. The major finding were as follows: (1) Delay time increased with age. (2) No sex difference is found in delay time. (3) Reflective children delayed longer than impulsive children in all age groups. (4) The reward which wasn't physically presented produced loner delay time than the reward which was physically presented in all age groups.

  • PDF

Scheduling Algorithms for the Maximal Total Revenue on a Single Processor with Starting Time Penalty

  • Joo, Un-Gi
    • Management Science and Financial Engineering
    • /
    • 제18권1호
    • /
    • pp.13-20
    • /
    • 2012
  • This paper considers a revenue maximization problem on a single processor. Each job is identified as its processing time, initial reward, reward decreasing rate, and preferred start time. If the processor starts a job at time zero, revenue of the job is its initial reward. However, the revenue decreases linearly with the reward decreasing rate according to its processing start time till its preferred start time and finally its revenue is zero if it is started the processing after the preferred time. Our objective is to find the optimal sequence which maximizes the total revenue. For the problem, we characterize the optimal solution properties and prove the NP-hardness. Based upon the characterization, we develop a branch-and-bound algorithm for the optimal sequence and suggest five heuristic algorithms for efficient solutions. The numerical tests show that the characterized properties are useful for effective and efficient algorithms.

Optimal Control Of Two-Hop Routing In Dtns With Time-Varying Selfish Behavior

  • Wu, Yahui;Deng, Su;Huang, Hongbin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제6권9호
    • /
    • pp.2202-2217
    • /
    • 2012
  • The transmission opportunities between nodes in Delay Tolerant Network (DTNs) are uncertain, and routing algorithms in DTNs often need nodes serving as relays for others to carry and forward messages. Due to selfishness, nodes may ask the source to pay a certain reward, and the reward may be varying with time. Moreover, the reward that the source obtains from the destination may also be varying with time. For example, the sooner the destination gets the message, the more rewards the source may obtain. The goal of this paper is to explore efficient ways for the source to maximize its total reward in such complex applications when it uses the probabilistic two-hop routing policy. We first propose a theoretical framework, which can be used to evaluate the total reward that the source can obtain. Then based on the model, we prove that the optimal forwarding policy confirms to the threshold form by the Pontryagin's Maximum Principle. Simulations based on both synthetic and real motion traces show the accuracy of our theoretical framework. Furthermore, we demonstrate that the performance of the optimal forwarding policy with threshold form is better through extensive numerical results, which conforms to the result obtained by the Maximum Principle.

STAD학습에서 복합보상이 학업성취도와 학습태도에 미치는 효과 (The Effect of the Complex Reward in STAD Learning on Academic Achievement and Learning Attitudes)

  • 김선수;최도성
    • 한국초등과학교육학회지:초등과학교육
    • /
    • 제21권1호
    • /
    • pp.101-109
    • /
    • 2002
  • A cooperative teaming has been taken to consolidate the autonomous motivation of students and to develop a desirable attitude in a mutual cooperative atmosphere. Some studies on the reward effect showed that the reward after the evaluation, in the processes of cooperative learning, worked on students' learning motive directly, and the group reward was effective in learning attitude and the individual reward in academic achievement, respectively. Assuming that the group reward and the individual reward are organized and applied as a complex reward, the effects of rewards will appear, this study examined the effect of the complex reward on academic achievement and teaming attitude. For this study. 2 classes were randomly selected out of a elementary school in Gwangju and the teaming unit was based on chapter 4「The structure and function of plants」 in the 5-1 elementary Science textbook. This research has been done for 4 weeks after the students learned STAD for 8 weeks previously. The learning attitude was examined in pre and post tests, and the academic achievement was inspected twice at 2-week intervals after the pre test. The results were analysized by the SAS program In the case of academic achievement, both groups showed a significant improvement(p<.05). The experimental group showed no significant improvement in the first test, compared with the control group(p>.05), but after 4 weeks, it showed a significant improvement in the second test, compared with the control group(p<.05). From this result, it is identified that the reward should be done for a long time and the individual reward of the complex reward is successful in improving academic achievement. However, in the case of learning attitude, there was no meaningful difference in both groups(p>.05). But the control group showed a significant improvement, compared with the experimental group(p<.05). According to this result, it is indicated that the group reward only is more effective in improving learning attitude and complex reward can decrease the individual competition in experimental group.

  • PDF

A Note on Renewal Reward Process with Fuzzy Rewards

  • Hong, Dug-Hun;Kim, Jeong-Jin;Do, Hae-Young
    • Journal of the Korean Data and Information Science Society
    • /
    • 제16권1호
    • /
    • pp.165-172
    • /
    • 2005
  • In recently, Popova and Wu(1999) proved a theorem which presents the long-run average fuzzy reward per unit time. In this note, we improve this result. Indeed we will show uniform convergence of a renewal reward processes with respect to the level ${\alpha}$ modeled as a fuzzy random variables.

  • PDF

스마트 제어알고리즘 개발을 위한 강화학습 리워드 설계 (Reward Design of Reinforcement Learning for Development of Smart Control Algorithm)

  • 김현수;윤기용
    • 한국공간구조학회논문집
    • /
    • 제22권2호
    • /
    • pp.39-46
    • /
    • 2022
  • Recently, machine learning is widely used to solve optimization problems in various engineering fields. In this study, machine learning is applied to development of a control algorithm for a smart control device for reduction of seismic responses. For this purpose, Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm. A single degree of freedom (SDOF) structure with a smart tuned mass damper (TMD) was used as an example structure. A smart TMD system was composed of MR (magnetorheological) damper instead of passive damper. Reward design of reinforcement learning mainly affects the control performance of the smart TMD. Various hyper-parameters were investigated to optimize the control performance of DQN-based control algorithm. Usually, decrease of the time step for numerical simulation is desirable to increase the accuracy of simulation results. However, the numerical simulation results presented that decrease of the time step for reward calculation might decrease the control performance of DQN-based control algorithm. Therefore, a proper time step for reward calculation should be selected in a DQN training process.

부분 해를 이용한 IRIS 실시간 태스크용 온-라인 스케줄링 알고리즘의 성능향상 (Performance Enhancement of On-Line Scheduling Algorithm for IRIS Real-Time Tasks using Partial Solution)

  • 심재홍;최경희;정기현
    • 한국정보과학회논문지:시스템및이론
    • /
    • 제30권1호
    • /
    • pp.12-21
    • /
    • 2003
  • 본 논문에서는 가치함수를 가지면서 동적으로 도착하는 IRIS(Increasing Reward with Increasing Service) 실시간 태스크들의 총 가치를 최대화하기 위한 온-라인 스케줄링 알고리즘을 제안한다. 본 논문은 스케줄링 알고리즘의 성능향상에 역점을 두고 있으며, 이는 다음 두 가지 아이디어를 기반으로 한다. 첫째, 총가치를 최대화하는 문제는 가치함수들의 최대 도함수 값들 중 최소 값을 찾는 문제를 해결함으로써 풀 수 있다는 것이다. 둘째, 새로운 태스크가 도착하기 전까지 이 전에 스케줄된 태스크들 중 소수만이 실제 실행되고, 나머지는 새로 도착한 태스크와 함께 다시 스케줄링 된다는 사실을 발견하고, 매 스케줄링 시 모든 태스크들을 스케줄링하는 것이 아니라, 일부 태스크들만 스케줄링하자는 것이다. 제안 알고리즘의 성능은 다양한 경우에 대한 모의실험으로 검증되었다. 실험 결과 제안 알고리즘의 계산 복잡도는 최악의 경우 기존 알고리즘과 동일한 $O(N_2)$이지만, 평균적으로 이 보다 낮은 O(N)에 가까운 것으로 확인되었다.

Exploring reward efficacy in traffic management using deep reinforcement learning in intelligent transportation system

  • Paul, Ananya;Mitra, Sulata
    • ETRI Journal
    • /
    • 제44권2호
    • /
    • pp.194-207
    • /
    • 2022
  • In the last decade, substantial progress has been achieved in intelligent traffic control technologies to overcome consistent difficulties of traffic congestion and its adverse effect on smart cities. Edge computing is one such advanced progress facilitating real-time data transmission among vehicles and roadside units to mitigate congestion. An edge computing-based deep reinforcement learning system is demonstrated in this study that appropriately designs a multiobjective reward function for optimizing different objectives. The system seeks to overcome the challenge of evaluating actions with a simple numerical reward. The selection of reward functions has a significant impact on agents' ability to acquire the ideal behavior for managing multiple traffic signals in a large-scale road network. To ascertain effective reward functions, the agent is trained withusing the proximal policy optimization method in several deep neural network models, including the state-of-the-art transformer network. The system is verified using both hypothetical scenarios and real-world traffic maps. The comprehensive simulation outcomes demonstrate the potency of the suggested reward functions.