A two-option reinforcement-learning task that measures how quickly participants track reward probabilities and update choices.
Configure parameters and run an interactive preview exactly as participants will experience it. No data is recorded.
Adjust parameters below, then start the preview on the right.
Include practice trials
Shown with feedback before the main task
Task parameters
Outputs: win-stay rate, lose-shift rate, arm choice proportion.
This is a researcher preview. No participant data is recorded.
Simulated participant view
A two-option reinforcement-learning task that measures how quickly participants track reward probabilities and update choices.
No data is recorded
Two options are shown on each trial; the participant chooses one and receives win or no-win feedback before the next trial.
Useful in studies of impulsivity, substance use, motivation, mood, and individual differences in reward sensitivity, where trial-by-trial learning dynamics are a target outcome.
Not ideal when the study needs classical cognitive measures such as memory or attention rather than learning-rate or value-based choice data.
Enable the optional reversal point to test flexible updating as well as initial acquisition. Keep the session short enough to avoid fatigue effects on learning.
Win-stay and lose-shift rates describe qualitative strategy use; fitting formal RL models (e.g., Rescorla-Wagner) to the trial sequence yields learning-rate and inverse-temperature estimates that are more sensitive.