Author: | Thomas V. Wiecki, Imri Sofer, Mads L. Pedersen, Alexander Fengler, Lakshmi Govindarajan, Krishn Bera, Michael J. Frank |
---|---|
Contact: | thomas.wiecki@gmail.com, imri_sofer@brown.edu, madslupe@gmail.com, alexander_fengler@brown.edu, krishn_bera@brown.edu, michael_frank@brown.edu |
Web site: | https://hddm.readthedocs.io |
Github: | http://github.com/hddm-devs/hddm |
Mailing list: | https://groups.google.com/group/hddm-users/ |
Copyright: | This document has been placed in the public domain. |
License: | HDDM is released under the BSD 2 license. |
Version: | 1.0.1 |
HDDM is a python toolbox for hierarchical Bayesian parameter estimation of the Drift Diffusion Model (via PyMC). Drift Diffusion Models are used widely in psychology and cognitive neuroscience to study decision making.
Check out the tutorial on how to get started. Further information can be found below as well as in the howto section and the documentation.
Uses hierarchical Bayesian estimation (via PyMC) of DDM parameters to allow simultaneous estimation of subject and group parameters, where individual subjects are assumed to be drawn from a group distribution. HDDM should thus produce better estimates when less RT values are measured compared to other methods using maximum likelihood for individual subjects (i.e. DMAT or fast-dm).
Heavily optimized likelihood functions for speed (Navarro & Fuss, 2009).
Flexible creation of complex models tailored to specific hypotheses (e.g. estimation of separate drift-rates for different task conditions; or predicted changes in model parameters as a function of other indicators like brain activity).
Estimate trial-by-trial correlations between a brain measure (e.g. fMRI BOLD) and a diffusion model parameter using the HDDMRegression model.
Built-in Bayesian hypothesis testing and several convergence and goodness-of-fit diagnostics.
As of version 0.7.1 HDDM includes modules for analyzing reinforcement learning data with the reinforcement learning drift diffusion model (RLDDM), including a module for estimating the impact of continuous regressors onto RLDDM parameters, and a reinforcement learning (RL) model. See tutorial for the RLDDM and RL modules here: https://nbviewer.jupyter.org/github/hddm-devs/hddm/blob/master/hddm/examples/demo_RLHDDMtutorial.ipynb and in the paper here: https://rdcu.be/b4q6Z
- HDDM 0.9.0 brings a host of new features.
HDDM includes likelihood approximation networks via the HDDMnn, HDDMnnRegressor and HDDMnnStimCoding classes. This allows fitting of a number of variants of sequential sampling models. You can now easily use custom likelihoods for model fitting. We included a range of new simulators, which allow data generation for a host of variants of sequential sampling models. There are some new out of the box plots, in the hddm.plotting module. Fast posterior predictives for regression based models. Some sampler settings are now exposed to the user via a customizable model_config dictionary. Lastly you are now able to save and load HDDMRegression models with custom link functions. Please see the documentation (under LAN Extension) for illustrations on how to use the new features.
- HDDM 0.9.1 improved documentation for LAN models.
Comprehensive tutorial using LAN included. Bugfixes for
simulator_h_c()
function.
- HDDM 0.9.2 major overhaul of the plotting functions under hddm.plotting.
Old capabilities are preserved under
hddm.plotting_old
, but will be deprecated. The new plotting functions replicate the existing functionality, but improve on various aspects of the plot and provide a more abstracted and extensible interface. Fixes an error with posterior predictive sampling using hierarchical regression models based on LANs withHDDMnnRegressor()
.HDDMnnRegressor()
now issues a single warning for boundary condition violations instead of flagging all occurences.
- HDDM 0.9.3 Lot's of minor improvements.
Model plots now working for race and lca models with n > 2 choices (use _plot_func_model_n as plot_func argument in hddm.plotting.plot_posterior_predictive). model_config files are simplified and class construction is a bit more robust to lack of specification, improving ease of use with custom models. Various plots received a bit more styling features. Better defaults for simulator_h_c function in hddm.simulators.hddm_dataset_generators. Posterior predictives now properly take into account the p_outlier parameter when generating data from the implicit mixture model. p_outlier percent of the data, now explicitly come from random choices with uniform reaction times. The documentation is updated to reflect / illustrate the improvements.
- HDDM 0.9.4 Bug fixes and one major new functionality.
HDDMnnRegressor now allows you to define indirect regressors, latent parameters that are driven by their own regression and link to model parameters. See the documentation for more information on this. Note this functionality is experimental for now. Model fitting will work, but extraenous functionality may not, including posterior predictives for models that include such indirect regressors. Including indirect regressors might demand you to think carefully about the supplied model_config. E.g. in the race_no_bias_3 model, the usual lower bounds on the v0, v1 ,v2, v3 parameters are 0. If we allow these parameters to be driven by an indirect regressor v, which is added to the regressions of v0, v1, v2, v3, then v0, v1, v2, v3
- HDDM 0.9.5 Bug fixes and another new functionality.
HDDMnnRegressor now allows you to also define indirect betas, latent parameters that can be used in regression models. E.g. in the race_no_bias_3 model, you can define a v beta, which will be expressed in the regression models for v0, v1, v2 like so:
v0 = beta0 * covariate0 + ... + v * covariate_v_0
v1 = beta0 * covariate0 + ... + v * covariate_v_1
v2 = beta0 * covariate0 + ... + v * covariate_v_2
The v0, v1, v2 parameters might be drifts in a preference based choice task, and dedicated to respective choice options in a stimulus screen. You may be interested in making these slopes dependent on the respective values of these options in a given trial, but wish to use a central beta, (v in this example) that relates option value to drift for all drifts.
Note that first, this is complementary to the indirect regressors defined above, and second that the usual modelling caveats, such as potential convergence problems apply.
Note also that the usage of indirect betas as well as indirect regressors may affect the speed of sampling in general. Both translate into more computational work at the stage of regression likelihood evaluation.
- HDDM 0.9.6 brings a host of new features.
HDDM now includes use of likelihood approximation networks in conjunction with reinforcement learning models via the HDDMnnRL class. This allows researchers to study not only the across-trial dynamics of learning but the within-trial dynamics of choice processes, using a single model. This module greatly extends the previous functionality for fitting RL+DDM models (via HDDMrl class) by allowing fitting of a number of variants of sequential sampling models in conjuction with a learning process (RL+SSM models). We have included a new simulator, which allows data generation for a host of variants of sequential sampling models in conjunction with the Rescorla-Wagner update rule on a 2-armed bandit task environment. There are some new, out-of-the-box plots and utility function in the hddm.plotting and hddm.utils modules, respectively, to facilitate posterior visualization and posterior predictive checks. Lastly you can also save and load HDDMnnRL models. Please see the documentation (under HDDMnnRL Extension) for illustrations on how to use the new features.
- HDDM 0.9.7 adds the HDDMnnRLRegressor class, the equivalent to the HDDMrlRegressor with support for many more SSMs via LANs.
Please check the documentation for usage examples.
- HDDM 0.9.8 adds a breaking change.
To accommodate some user requests, all parameters that should be estimated for a given model should be made explicitly in the include argument of the calling class (HDDM, HDDMnn, etc.). The remaining parameters can now be set to arbitrary, user defined, defaults. Check the documentation for a new tutorial on parameter defaults. Moreover, model plot is made much more flexible (new tutorial included to showcase some of the options). Two tutorials are added to showcase the capabilities for simulation based inference via custom likelihoods. The legacy models with "vanilla" in their name are globally renamed to instead include "hddm_base". The simulator backend is now completely outsourced to the ssms package (severe code simplifications).
A recent paper by Roger Ratcliff quantitatively compared DMAT, fast-dm, and EZ, and concluded: "We found that the hierarchical diffusion method [as implemented by HDDM] performed very well, and is the method of choice when the number of observations is small."
Find the paper here: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4517692/
The following is a minimal python script to load data, run a model and examine its parameters and fit.
import hddm # Load data from csv file into a NumPy structured array data = hddm.load_csv('simple_difficulty.csv') # Create a HDDM model multi object model = hddm.HDDM(data, depends_on={'v':'difficulty'}) # Create model and start MCMC sampling model.sample(2000, burn=20) # Print fitted parameters and other model statistics model.print_stats() # Plot posterior distributions and theoretical RT distributions model.plot_posteriors() model.plot_posterior_predictive()
For more information about the software and theories behind it, please see the main publication.
For HDDM >= 0.9.0, currently in beta release, the most convenient way to install HDDM, is to directly install via github. In a fresh environment (we recommend to use python 3.7) type.
We recommend you to open a conda environment first.
conda create -n hddm python=3.7 conda activate hddm
If you do not have hdf5 or netcdf4 installed, you can use conda to install them.
conda install -c conda-forge hdf5 conda install -c conda-forge netcdf4
Then install hddm via github.
pip install "numpy < 1.23.0" pip install "cython <1.0.0" pip install pymc==2.3.8 # backend probabilistic programming framework (DO NOT USE CONDA HERE) # pip install git+https://github.com/hddm-devs/kabuki # (Optional, the pip version as dependency of HDDM should work) pip install hddm # main package # Optional pip install torch torchvision torchaudio # The LAN extension makes use of these
To make use of the LAN fuctionalities, you need to install pytorch.
- ::
- pip install torch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1
A common issue on new machines is that the installation of the pymc package (a necessary dependency), is hampered by problems with compiling its fortran code. Try downgrading the version of your gcc compiler. This can be done on a MAC (not the new M1/M2 versions tragically), via
brew install gcc@9
In case you do not have the brew command, install Homebrew first.
You usually do not run into problems with linux machines, however downgrading gcc can still be necessary.
(Previous instructions for HDDM <= 0.8.0, DISCOURAGED) As of release 0.6.0, HDDM is compatible with Python 3 which we encourage.
The easiest way to install HDDM is through Anaconda (available for Windows, Linux and OSX):
- Download and install Anaconda.
- In a shell (Windows: Go to Start->Programs->Anaconda->Anaconda command prompt) type:
conda install -c pymc hddm
If you want to use pip instead of conda, type:
pip install pandas pip install pymc==2.3.8 pip install kabuki pip install hddm
This might require super-user rights via sudo. Note that this installation method is discouraged as it leads to all kinds of problems on various platforms.
If you are having installation problems please contact the mailing list.
And if you're a mac user, check out this thread for advice on installation.
If HDDM was used in your research, please cite the main publication:
Wiecki TV, Sofer I and Frank MJ (2013). HDDM: Hierarchical Bayesian estimation of the Drift-Diffusion Model in Python. Front. Neuroinform. 7:14. doi: 10.3389/fninf.2013.00014
If you use the HDDMrl, please cite the original HDDM RL tutorial paper:
Pedersen, M. L., & Frank, M. J. (2020). Simultaneous hierarchical bayesian parameter estimation for reinforcement learning and drift diffusion models: a tutorial and links to neural data. Computational Brain & Behavior, 3(4), 458-471.
If you use any of the HDDMnn, HDDMnnRegressor, HDDMnnStimCoding or HDDMnnRL classes, please cite the lan extension and the new tutorial paper:
Alexander Fengler, Lakshmi N Govindarajan, Tony Chen, Michael J Frank (2021). Likelihood approximation networks (LANs) for fast inference of simulation models in cognitive neuroscience. eLife 10:e65074. doi: 10.7554/eLife.65074
Fengler, A., Bera, K., Pedersen, M. L., & Frank, M. J. (2022). Beyond Drift Diffusion Models: Fitting a Broad Class of Decision and Reinforcement Learning Models with HDDM. Journal of Cognitive Neuroscience, 34(10), 1780-1805.
HDDM has been used in over 400 published papers.
James Rowe (Cambridge University): "The HDDM modelling gave insights into the effects of disease that were simply not visible from a traditional analysis of RT/Accuracy. It provides a clue as to why many disorders including PD and PSP can give the paradoxical combination of akinesia and impulsivity. Perhaps of broader interest, the hierarchical drift diffusion model turned out to be very robust. In separate work, we have found that the HDDM gave accurate estimates of decision parameters with many fewer than 100 trials, in contrast to the hundreds or even thousands one might use for ‘traditional’ DDMs. This meant it was realistic to study patients who do not tolerate long testing sessions."
Check out the tutorial on how to get started. Further information can be found in howto and the documentation.
Join our low-traffic mailing list.