NeuroGym¶
Tasks¶
Currently implemented tasks can be found here.
Wrappers¶
Wrappers (see here) are short scripts that allow introducing modifications the original tasks. For instance, the Random Dots Motion task can be transformed into a reaction time task by passing it through the reaction_time wrapper. Alternatively, the combine wrapper allows training an agent in two different tasks simultaneously.
Configuration¶
🧪 Beta Feature — The configuration system is optional and currently under development. You can still instantiate environments, agents, and wrappers with direct parameters.
It is only used in a small portion of the codebase and is not required for typical usage.
See the demo.ipynb notebook for the only current example of this system in action.
NeuroGym includes a flexible configuration mechanism using Pydantic Settings, allowing configuration via TOML files, Python objects, or plain dictionaries.
Using a TOML file can be especially useful for sharing experiment configurations in a portable way (e.g., sending config.toml to a colleague), reliably saving and loading experiment setups, and easily switching between multiple configurations for the same environment by changing just one line of code. While the system isn't at that stage yet, these are intended future capabilities.
1. From a TOML file¶
Create a config.toml file (see template) and load it:
You can then pass this config to any component that supports it:
from neurogym import make
from neurogym.wrappers import Monitor
env = make('GoNogo-v0')
env = Monitor(env, config=config)
Or directly pass the path:
2. With Python Class¶
from neurogym.config import Config
config = Config(
local_dir="logs/",
env={"name": "GoNogo-v0"},
monitor={"name": "MyMonitor"}
)
3. With a Dictionary¶
from neurogym.config import Config
config_dict = {
"env": {"name": "GoNogo-v0"},
"monitor": {
"name": "MyMonitor",
"plot": {"trigger": "step", "value": 500, "create": True}
},
"local_dir": "./outputs"
}
config = Config.model_validate(config_dict)
Examples¶
NeuroGym is compatible with most packages that use gymnasium. In this example jupyter notebook we show how to train a neural network with reinforcement learning algorithms using the Stable-Baselines3 toolbox.
Vanilla RNN Support in RecurrentPPO¶
We extended the RecurrentPPO implementation from stable-baselines3-contrib to support vanilla RNNs (torch.nn.RNN) in addition to LSTMs. This is particularly useful for neuroscience applications, where simpler recurrent architectures can be more biologically interpretable.
You can enable vanilla RNNs by setting recurrent_layer_type="rnn" in the policy_kwargs:
from sb3_contrib import RecurrentPPO
policy_kwargs = {"recurrent_layer_type": "rnn"} # "lstm" is the default
model = RecurrentPPO("MlpLstmPolicy", env_vec, policy_kwargs=policy_kwargs, verbose=1)
model.learn(5000)
Note: This feature is part of an open pull request to the upstream repository and is currently under review by the maintainers. Until the pull request is merged, you can use this functionality by installing NeuroGym organization's fork of the repository. To do so, uninstall the original package and install from the custom branch:
pip uninstall stable-baselines3-contrib -y
pip install git+https://github.com/neurogym/stable-baselines3-contrib.git@rnn_policy_addition
This will install the version with vanilla RNN support from the rnn_policy_addition branch in our fork.
Custom Tasks¶
Creating custom new tasks should be easy. You can contribute tasks using the regular gymnasium format. If your task has a trial/period structure, this template provides the basic structure that we recommend a task to have:
from neurogym import spaces
import neurogym as ngym
class YourTask(ngym.PeriodEnv):
metadata = {}
def __init__(self, dt=100, timing=None, extra_input_param=None):
super().__init__(dt=dt)
def new_trial(self, **kwargs):
"""
new_trial() is called when a trial ends to generate the next trial.
Here you have to set:
The trial periods: fixation, stimulus...
Optionally, you can set:
The ground truth: the correct answer for the created trial.
"""
def _step(self, action):
"""
_step receives an action and returns:
a new observation, obs
reward associated with the action, reward
a boolean variable indicating whether the experiment has terminated, terminated
See more at https://gymnasium.farama.org/tutorials/gymnasium_basics/handling_time_limits/#termination
a boolean variable indicating whether the experiment has been truncated, truncated
See more at https://gymnasium.farama.org/tutorials/gymnasium_basics/handling_time_limits/#truncation
a dictionary with extra information:
ground truth correct response, info['gt']
boolean indicating the end of the trial, info['new_trial']
"""
return obs, reward, terminated, truncated, {'new_trial': new_trial, 'gt': gt}