• Title/Summary/Keyword: Reward Probability Processing

Search Result 3, Processing Time 0.022 seconds

Estimation of Reward Probability in the Fronto-parietal Functional Network: An fMRI Study

  • Shin, Yeonsoon;Kim, Hye-young;Min, Seokyoung;Han, Sanghoon
    • Science of Emotion and Sensibility
    • /
    • v.20 no.4
    • /
    • pp.101-112
    • /
    • 2017
  • We investigated the neural representation of reward probability recognition and its neural connectivity with other regions of the brain. Using functional magnetic resonance imaging (fMRI), we used a simple guessing task with different probabilities of obtaining rewards across trials to assay local and global regions processing reward probability. The results of whole brain analysis demonstrated that lateral prefrontal cortex, inferior parietal lobe, and postcentral gyrus were activated during probability-based decision making. Specifically, the higher the expected value was, the more these regions were activated. Fronto-parietal connectivity, comprising inferior parietal regions and right lateral prefrontal cortex, conjointly engaged during high reward probability recognition compared to low reward condition, regardless of whether the reward information was extrinsically presented. Finally, the result of a regression analysis identified that cortico-subcortical connectivity was strengthened during the high reward anticipation for the subjects with higher cognitive impulsivity. Our findings demonstrate that interregional functional involvement is involved in valuation based on reward probability and that personality trait such as cognitive impulsivity plays a role in modulating the connectivity among different brain regions.

Evaluating SR-Based Reinforcement Learning Algorithm Under the Highly Uncertain Decision Task (불확실성이 높은 의사결정 환경에서 SR 기반 강화학습 알고리즘의 성능 분석)

  • Kim, So Hyeon;Lee, Jee Hang
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.8
    • /
    • pp.331-338
    • /
    • 2022
  • Successor representation (SR) is a model of human reinforcement learning (RL) mimicking the underlying mechanism of hippocampal cells constructing cognitive maps. SR utilizes these learned features to adaptively respond to the frequent reward changes. In this paper, we evaluated the performance of SR under the context where changes in latent variables of environments trigger the reward structure changes. For a benchmark test, we adopted SR-Dyna, an integration of SR into goal-driven Dyna RL algorithm in the 2-stage Markov Decision Task (MDT) in which we can intentionally manipulate the latent variables - state transition uncertainty and goal-condition. To precisely investigate the characteristics of SR, we conducted the experiments while controlling each latent variable that affects the changes in reward structure. Evaluation results showed that SR-Dyna could learn to respond to the reward changes in relation to the changes in latent variables, but could not learn rapidly in that situation. This brings about the necessity to build more robust RL models that can rapidly learn to respond to the frequent changes in the environment in which latent variables and reward structure change at the same time.

A Novel Smart Contract based Optimized Cloud Selection Framework for Efficient Multi-Party Computation

  • Haotian Chen;Abir EL Azzaoui;Sekione Reward Jeremiah;Jong Hyuk Park
    • Journal of Information Processing Systems
    • /
    • v.19 no.2
    • /
    • pp.240-257
    • /
    • 2023
  • The industrial Internet of Things (IIoT) is characterized by intelligent connection, real-time data processing, collaborative monitoring, and automatic information processing. The heterogeneous IIoT devices require a high data rate, high reliability, high coverage, and low delay, thus posing a significant challenge to information security. High-performance edge and cloud servers are a good backup solution for IIoT devices with limited capabilities. However, privacy leakage and network attack cases may occur in heterogeneous IIoT environments. Cloud-based multi-party computing is a reliable privacy-protecting technology that encourages multiparty participation in joint computing without privacy disclosure. However, the default cloud selection method does not meet the heterogeneous IIoT requirements. The server can be dishonest, significantly increasing the probability of multi-party computation failure or inefficiency. This paper proposes a blockchain and smart contract-based optimized cloud node selection framework. Different participants choose the best server that meets their performance demands, considering the communication delay. Smart contracts provide a progressive request mechanism to increase participation. The simulation results show that our framework improves overall multi-party computing efficiency by up to 44.73%.