Decision Making3–5 minBurden: MediumEMA: Medium

Probabilistic Learning

A two-option reinforcement-learning task that measures how quickly participants track reward probabilities and update choices.

Reinforcement learningDecision makingReward sensitivity
Category
Decision Making
Typical duration
3–5 min
Participant burden
Medium
EMA suitability
Medium

Try this task

Configure parameters and run an interactive preview exactly as participants will experience it. No data is recorded.

Configure preview

Adjust parameters below, then start the preview on the right.

Include practice trials

Shown with feedback before the main task

Task parameters

Outputs: win-stay rate, lose-shift rate, arm choice proportion.

3–5 minBurden: MediumEMA: Medium

This is a researcher preview. No participant data is recorded.

Simulated participant view

9:41

Probabilistic Learning

A two-option reinforcement-learning task that measures how quickly participants track reward probabilities and update choices.

No data is recorded

Participant experience on smartphone

Two options are shown on each trial; the participant chooses one and receives win or no-win feedback before the next trial.

When to use

Useful in studies of impulsivity, substance use, motivation, mood, and individual differences in reward sensitivity, where trial-by-trial learning dynamics are a target outcome.

When not to use

Not ideal when the study needs classical cognitive measures such as memory or attention rather than learning-rate or value-based choice data.

How to use in a study

Enable the optional reversal point to test flexible updating as well as initial acquisition. Keep the session short enough to avoid fatigue effects on learning.

Researcher-configurable parameters

  • Number of trials
  • Reward probability for each arm
  • Optional reversal trial index
  • Feedback display duration
  • Practice block enabled / disabled

Outputs collected

  • Choice proportion per arm
  • Win-stay rate
  • Lose-shift rate
  • Post-reversal accuracy (if reversal enabled)

Interpretation notes

Win-stay and lose-shift rates describe qualitative strategy use; fitting formal RL models (e.g., Rescorla-Wagner) to the trial sequence yields learning-rate and inverse-temperature estimates that are more sensitive.

Scientific evidence

  • Smartphone probabilistic-learning tasks recover meaningful individual differences in reinforcement learning parameters.
  • Win-stay / lose-shift metrics are interpretable without model fitting and useful for exploratory analyses.

Read more

Related tasks

Use in your study

Add this task to your ESM study on SMAAT — free to get started.

Sign up free →